2025-04-02 02:17:45,031 [ 51330 ] INFO : ClickHouse root is not set. Will use /home/ubuntu/_work/ClickHouse/ClickHouse (runner:53, check_args_and_update_paths) 2025-04-02 02:17:45,031 [ 51330 ] INFO : Cases dir is not set. Will use /home/ubuntu/_work/ClickHouse/ClickHouse/tests/integration (runner:97, check_args_and_update_paths) 2025-04-02 02:17:45,031 [ 51330 ] INFO : utils dir is not set. Will use /home/ubuntu/_work/ClickHouse/ClickHouse/utils (runner:108, check_args_and_update_paths) 2025-04-02 02:17:45,031 [ 51330 ] INFO : base_configs_dir: /home/ubuntu/_work/ClickHouse/ClickHouse/programs/server, binary: /home/ubuntu/_work/_temp/test/build/clickhouse, cases_dir: /home/ubuntu/_work/ClickHouse/ClickHouse/tests/integration (runner:110, check_args_and_update_paths) clickhouse_integration_tests_volume Running pytest container as: 'docker run --rm --name clickhouse_integration_tests_4gikvv --privileged --dns-search='.' --memory=30709030912 --security-opt seccomp=unconfined --cap-add=SYS_PTRACE --volume=/home/ubuntu/_work/_temp/test/build/clickhouse-odbc-bridge:/clickhouse-odbc-bridge --volume=/home/ubuntu/_work/_temp/test/build/clickhouse:/clickhouse --volume=/home/ubuntu/_work/_temp/test/build/clickhouse-library-bridge:/clickhouse-library-bridge --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/programs/server:/clickhouse-config --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/tests/integration:/ClickHouse/tests/integration --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/utils/backupview:/ClickHouse/utils/backupview --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/utils/grpc-client/pb2:/ClickHouse/utils/grpc-client/pb2 --volume=/run:/run/host:ro --volume=clickhouse_integration_tests_volume:/var/lib/docker -e DOCKER_DOTNET_CLIENT_TAG=11de0b29a15d -e DOCKER_HELPER_TAG=5dc43a6382f0 -e DOCKER_BASE_TAG=8b2301119731 -e DOCKER_KERBEROS_KDC_TAG=9391ecdee8d7 -e DOCKER_MYSQL_GOLANG_CLIENT_TAG=9bec2a638e6e -e DOCKER_MYSQL_JAVA_CLIENT_TAG=766bff31cfe4 -e DOCKER_MYSQL_JS_CLIENT_TAG=41ba7c2ec2a1 -e DOCKER_MYSQL_PHP_CLIENT_TAG=88be89c1e3b6 -e DOCKER_NGINX_DAV_TAG=b55ac9cd7519 -e DOCKER_POSTGRESQL_JAVA_CLIENT_TAG=a4eff5c7f4d6 -e DOCKER_PYTHON_BOTTLE_TAG=caad4729259e -e DOCKER_CLIENT_TIMEOUT=300 -e COMPOSE_HTTP_TIMEOUT=600 -e PYTHONUNBUFFERED=1 -e PYTEST_ADDOPTS="--dist=loadfile -n 10 -rfEps --run-id=0 --color=no --durations=0 test_aggregation_memory_efficient/test.py::test_remote test_allow_feature_tier/test.py::test_allow_feature_tier_in_general_settings test_allow_feature_tier/test.py::test_allow_feature_tier_in_mergetree_settings test_allow_feature_tier/test.py::test_allow_feature_tier_in_mergetree_settings_with_old_compatibility test_allow_feature_tier/test.py::test_allow_feature_tier_in_user test_allow_feature_tier/test.py::test_it_is_possible_to_enable_experimental_settings_in_default_profile test_analyzer_compatibility/test.py::test_two_new_versions test_async_load_databases/test.py::test_async_load_system_database test_async_load_databases/test.py::test_dependent_tables test_async_load_databases/test.py::test_dict_get_data test_async_load_databases/test.py::test_multiple_tables test_attach_table_normalizer/test.py::test_attach_substr test_attach_table_normalizer/test.py::test_attach_substr_restart test_azure_blob_storage_zero_copy_replication/test.py::test_zero_copy_replication test_backup_restore/test.py::test_attach_partition test_backup_restore/test.py::test_replace_partition test_backup_restore/test.py::test_restore 'test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[http-False]' 'test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[http-True]' 'test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[native-True]' test_backup_restore_on_cluster/test.py::test_backup_restore_on_single_replica test_backup_restore_on_cluster/test.py::test_different_tables_on_nodes test_backup_restore_on_cluster/test.py::test_empty_replicated_table test_backup_restore_on_cluster/test.py::test_file_deduplication test_backup_restore_on_cluster/test.py::test_get_error_from_other_host test_backup_restore_on_cluster/test.py::test_keeper_value_max_size test_backup_restore_on_cluster/test.py::test_mutation test_backup_restore_on_cluster/test.py::test_projection test_backup_restore_on_cluster/test.py::test_replicated_database test_backup_restore_on_cluster/test.py::test_replicated_database_async test_backup_restore_on_cluster/test.py::test_replicated_database_compare_parts 'test_backup_restore_on_cluster/test.py::test_replicated_database_with_special_macro_in_zk_path[database]' 'test_backup_restore_on_cluster/test.py::test_replicated_database_with_special_macro_in_zk_path[uuid]' test_backup_restore_on_cluster/test.py::test_replicated_table test_backup_restore_on_cluster/test.py::test_replicated_table_after_alters test_backup_restore_on_cluster/test.py::test_replicated_table_restored_into_bigger_cluster test_backup_restore_on_cluster/test.py::test_replicated_table_restored_into_smaller_cluster test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_def test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_insert test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_merge test_backup_restore_on_cluster/test.py::test_replicated_table_with_uuid_in_zkpath test_backup_restore_on_cluster/test.py::test_required_privileges test_backup_restore_on_cluster/test.py::test_shutdown_waits_for_backup test_backup_restore_on_cluster/test.py::test_system_functions test_backup_restore_on_cluster/test.py::test_system_users test_backup_restore_on_cluster/test.py::test_table_in_replicated_database_with_not_synced_def test_backup_restore_on_cluster/test.py::test_table_with_parts_in_queue_considered_non_empty test_backup_restore_on_cluster/test.py::test_tables_dependency test_cluster_all_replicas/test.py::test_cluster 'test_cluster_all_replicas/test.py::test_error_on_unavailable_replica[one_shard_three_nodes]' 'test_cluster_all_replicas/test.py::test_error_on_unavailable_replica[two_shards_three_nodes]' test_cluster_all_replicas/test.py::test_global_in 'test_cluster_all_replicas/test.py::test_skip_unavailable_replica[one_shard_three_nodes]' 'test_cluster_all_replicas/test.py::test_skip_unavailable_replica[two_shards_three_nodes]' test_composable_protocols/test.py::test_connections test_composable_protocols/test.py::test_http_proxy_1 test_composable_protocols/test.py::test_proxy_1 test_config_corresponding_root/test.py::test_work test_config_decryption/test.py::test_successful_decryption_xml test_config_decryption/test.py::test_successful_decryption_yaml test_config_hide_in_preprocessed/test.py::test_hide_in_preprocessed test_config_substitutions/test.py::test_allow_databases test_config_substitutions/test.py::test_config test_config_substitutions/test.py::test_config_from_env_overrides test_config_substitutions/test.py::test_config_merge_from_env_overrides test_config_substitutions/test.py::test_config_multiple_zk_substitutions test_config_substitutions/test.py::test_include_config 'test_database_backup/test.py::test_database_backup_database[Disk('\"'\"'backup_disk_local'\"'\"', '\"'\"'test_database_backup'\"'\"')]' 'test_database_backup/test.py::test_database_backup_database[Disk('\"'\"'backup_disk_object_storage_local_plain'\"'\"', '\"'\"'test_database_backup'\"'\"')]' 'test_database_backup/test.py::test_database_backup_database[Disk('\"'\"'backup_disk_s3_plain'\"'\"', '\"'\"'test_database_backup'\"'\"')]' 'test_database_backup/test.py::test_database_backup_database[File('\"'\"'test_database_backup_file'\"'\"')]' 'test_database_backup/test.py::test_database_backup_table[Disk('\"'\"'backup_disk_local'\"'\"', '\"'\"'test_table_backup'\"'\"')]' 'test_database_backup/test.py::test_database_backup_table[Disk('\"'\"'backup_disk_object_storage_local_plain'\"'\"', '\"'\"'test_table_backup'\"'\"')]' 'test_database_backup/test.py::test_database_backup_table[Disk('\"'\"'backup_disk_s3_plain'\"'\"', '\"'\"'test_table_backup'\"'\"')]' 'test_database_backup/test.py::test_database_backup_table[File('\"'\"'test_table_backup_file'\"'\"')]' 'test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_complex[complex_key_cache]' 'test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_complex[complex_key_direct]' 'test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_complex[complex_key_hashed]' 'test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_ranged[range_hashed]' 'test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_simple[cache]' 'test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_simple[direct]' 'test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_simple[flat]' 'test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_simple[hashed]' 'test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_complex[complex_key_cache]' 'test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_complex[complex_key_direct]' 'test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_complex[complex_key_hashed]' 'test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_ranged[range_hashed]' 'test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_simple[cache]' 'test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_simple[direct]' 'test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_simple[flat]' 'test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_simple[hashed]' 'test_dictionaries_redis/test.py::test_redis_dictionaries[DateTime_]' 'test_dictionaries_redis/test.py::test_redis_dictionaries[Date_]' 'test_dictionaries_redis/test.py::test_redis_dictionaries[Float32_]' 'test_dictionaries_redis/test.py::test_redis_dictionaries[Float64_]' 'test_dictionaries_redis/test.py::test_redis_dictionaries[Int16_]' 'test_dictionaries_redis/test.py::test_redis_dictionaries[Int32_]' 'test_dictionaries_redis/test.py::test_redis_dictionaries[Int64_]' 'test_dictionaries_redis/test.py::test_redis_dictionaries[Int8_]' 'test_dictionaries_redis/test.py::test_redis_dictionaries[String_]' -vvv -ss" altinityinfra/integration-tests-runner:2165613c5fcd '. Start tests ============================= test session starts ============================== platform linux -- Python 3.10.12, pytest-7.4.4, pluggy-1.5.0 -- /usr/bin/python3 cachedir: .pytest_cache Test order randomisation NOT enabled. Enable with --random-order or --random-order-bucket= rootdir: /ClickHouse/tests/integration configfile: pytest.ini plugins: timeout-2.3.1, repeat-0.9.3, order-1.0.0, reportlog-0.4.0, xdist-3.5.0, random-order-1.1.1 timeout: 900.0s timeout method: signal timeout func_only: False created: 10/10 workers 10 workers [100 items] scheduling tests via LoadFileScheduling test_dictionaries_redis/test.py::test_redis_dictionaries[DateTime_] test_database_backup/test.py::test_database_backup_database[Disk('backup_disk_local', 'test_database_backup')] test_config_substitutions/test.py::test_allow_databases Command:[docker ps | wc -l] Command:[docker ps | wc -l] Command:[docker ps | wc -l] Command:[docker ps | wc -l] Command:[docker ps | wc -l] Command:[docker ps | wc -l] Command:[docker ps | wc -l] Command:[docker ps | wc -l] Command:[docker ps | wc -l] test_allow_feature_tier/test.py::test_allow_feature_tier_in_general_settings Command:[docker ps | wc -l] test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_complex[complex_key_cache] test_cluster_all_replicas/test.py::test_cluster test_backup_restore/test.py::test_attach_partition test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[http-False] test_async_load_databases/test.py::test_async_load_system_database test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_complex[complex_key_cache] Stdout:1 No running containers Pruning Docker networks Command:[docker network prune --force] Stdout:1 No running containers Pruning Docker networks Command:[docker network prune --force] Stdout:1 No running containers Pruning Docker networks Command:[docker network prune --force] Stdout:1 No running containers Pruning Docker networks Command:[docker network prune --force] Stdout:1 Stdout:1 No running containers Pruning Docker networks Command:[docker network prune --force] No running containers Pruning Docker networks Command:[docker network prune --force] Stdout:1 Stdout:1 No running containers Stdout:1 No running containers No running containers Pruning Docker networks Pruning Docker networks Pruning Docker networks Command:[docker network prune --force] Command:[docker network prune --force] Command:[docker network prune --force] Stdout:1 No running containers Pruning Docker networks Command:[docker network prune --force] Command:[sysctl net.ipv4.ip_local_port_range='55000 65535'] Command:[sysctl net.ipv4.ip_local_port_range='55000 65535'] Command:[sysctl net.ipv4.ip_local_port_range='55000 65535'] Command:[sysctl net.ipv4.ip_local_port_range='55000 65535'] Command:[sysctl net.ipv4.ip_local_port_range='55000 65535'] Stdout:net.ipv4.ip_local_port_range = 55000 65535 Stdout:net.ipv4.ip_local_port_range = 55000 65535 Stdout:net.ipv4.ip_local_port_range = 55000 65535 Stdout:net.ipv4.ip_local_port_range = 55000 65535 Running tests in /ClickHouse/tests/integration/test_database_backup/test.py ENV DOCKER_KERBEROS_KDC_TAG 9391ecdee8d7 Cluster start called. is_up=False ENV CLICKHOUSE_TESTS_SERVER_BIN_PATH /clickhouse ENV MSAN_OPTIONS abort_on_error=1 poison_in_dtor=1 Running tests in /ClickHouse/tests/integration/test_async_load_databases/test.py ENV JAVA_TOOL_OPTIONS -Djdk.attach.allowAttachSelf=true Source RedisSimple incompatible with layout complex_key_hashed Cluster start called. is_up=False Source RedisSimple incompatible with layout complex_key_cache ENV TSAN_OPTIONS halt_on_error=1 abort_on_error=1 history_size=7 memory_limit_mb=46080 second_deadlock_stack=1 ENV HOSTNAME 1dd346380e80 Command:[sysctl net.ipv4.ip_local_port_range='55000 65535'] ENV SHLVL 0 Source RedisSimple incompatible with layout complex_key_direct Source RedisHash incompatible with layout flat Source RedisHash incompatible with layout hashed ENV HOME /root Source RedisHash incompatible with layout cache ENV OLDPWD / ENV DOCKER_HELPER_TAG 5dc43a6382f0 ENV PYTHONUNBUFFERED 1 Command:[sysctl net.ipv4.ip_local_port_range='55000 65535'] ENV DOCKER_PYTHON_BOTTLE_TAG caad4729259e Command:[sysctl net.ipv4.ip_local_port_range='55000 65535'] ENV UBSAN_OPTIONS print_stacktrace=1 Source RedisHash incompatible with layout direct ENV PYTEST_ADDOPTS --dist=loadfile -n 10 -rfEps --run-id=0 --color=no --durations=0 test_aggregation_memory_efficient/test.py::test_remote test_allow_feature_tier/test.py::test_allow_feature_tier_in_general_settings test_allow_feature_tier/test.py::test_allow_feature_tier_in_mergetree_settings test_allow_feature_tier/test.py::test_allow_feature_tier_in_mergetree_settings_with_old_compatibility test_allow_feature_tier/test.py::test_allow_feature_tier_in_user test_allow_feature_tier/test.py::test_it_is_possible_to_enable_experimental_settings_in_default_profile test_analyzer_compatibility/test.py::test_two_new_versions test_async_load_databases/test.py::test_async_load_system_database test_async_load_databases/test.py::test_dependent_tables test_async_load_databases/test.py::test_dict_get_data test_async_load_databases/test.py::test_multiple_tables test_attach_table_normalizer/test.py::test_attach_substr test_attach_table_normalizer/test.py::test_attach_substr_restart test_azure_blob_storage_zero_copy_replication/test.py::test_zero_copy_replication test_backup_restore/test.py::test_attach_partition test_backup_restore/test.py::test_replace_partition test_backup_restore/test.py::test_restore 'test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[http-False]' 'test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[http-True]' 'test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[native-True]' test_backup_restore_on_cluster/test.py::test_backup_restore_on_single_replica test_backup_restore_on_cluster/test.py::test_different_tables_on_nodes test_backup_restore_on_cluster/test.py::test_empty_replicated_table test_backup_restore_on_cluster/test.py::test_file_deduplication test_backup_restore_on_cluster/test.py::test_get_error_from_other_host test_backup_restore_on_cluster/test.py::test_keeper_value_max_size test_backup_restore_on_cluster/test.py::test_mutation test_backup_restore_on_cluster/test.py::test_projection test_backup_restore_on_cluster/test.py::test_replicated_database test_backup_restore_on_cluster/test.py::test_replicated_database_async test_backup_restore_on_cluster/test.py::test_replicated_database_compare_parts 'test_backup_restore_on_cluster/test.py::test_replicated_database_with_special_macro_in_zk_path[database]' 'test_backup_restore_on_cluster/test.py::test_replicated_database_with_special_macro_in_zk_path[uuid]' test_backup_restore_on_cluster/test.py::test_replicated_table test_backup_restore_on_cluster/test.py::test_replicated_table_after_alters test_backup_restore_on_cluster/test.py::test_replicated_table_restored_into_bigger_cluster test_backup_restore_on_cluster/test.py::test_replicated_table_restored_into_smaller_cluster test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_def test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_insert test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_merge test_backup_restore_on_cluster/test.py::test_replicated_table_with_uuid_in_zkpath test_backup_restore_on_cluster/test.py::test_required_privileges test_backup_restore_on_cluster/test.py::test_shutdown_waits_for_backup test_backup_restore_on_cluster/test.py::test_system_functions test_backup_restore_on_cluster/test.py::test_system_users test_backup_restore_on_cluster/test.py::test_table_in_replicated_database_with_not_synced_def test_backup_restore_on_cluster/test.py::test_table_with_parts_in_queue_considered_non_empty test_backup_restore_on_cluster/test.py::test_tables_dependency test_cluster_all_replicas/test.py::test_cluster 'test_cluster_all_replicas/test.py::test_error_on_unavailable_replica[one_shard_three_nodes]' 'test_cluster_all_replicas/test.py::test_error_on_unavailable_replica[two_shards_three_nodes]' test_cluster_all_replicas/test.py::test_global_in 'test_cluster_all_replicas/test.py::test_skip_unavailable_replica[one_shard_three_nodes]' 'test_cluster_all_replicas/test.py::test_skip_unavailable_replica[two_shards_three_nodes]' test_composable_protocols/test.py::test_connections test_composable_protocols/test.py::test_http_proxy_1 test_composable_protocols/test.py::test_proxy_1 test_config_corresponding_root/test.py::test_work test_config_decryption/test.py::test_successful_decryption_xml test_config_decryption/test.py::test_successful_decryption_yaml test_config_hide_in_preprocessed/test.py::test_hide_in_preprocessed test_config_substitutions/test.py::test_allow_databases test_config_substitutions/test.py::test_config test_config_substitutions/test.py::test_config_from_env_overrides test_config_substitutions/test.py::test_config_merge_from_env_overrides test_config_substitutions/test.py::test_config_multiple_zk_substitutions test_config_substitutions/test.py::test_include_config 'test_database_backup/test.py::test_database_backup_database[Disk('"'"'backup_disk_local'"'"', '"'"'test_database_backup'"'"')]' 'test_database_backup/test.py::test_database_backup_database[Disk('"'"'backup_disk_object_storage_local_plain'"'"', '"'"'test_database_backup'"'"')]' 'test_database_backup/test.py::test_database_backup_database[Disk('"'"'backup_disk_s3_plain'"'"', '"'"'test_database_backup'"'"')]' 'test_database_backup/test.py::test_database_backup_database[File('"'"'test_database_backup_file'"'"')]' 'test_database_backup/test.py::test_database_backup_table[Disk('"'"'backup_disk_local'"'"', '"'"'test_table_backup'"'"')]' 'test_database_backup/test.py::test_database_backup_table[Disk('"'"'backup_disk_object_storage_local_plain'"'"', '"'"'test_table_backup'"'"')]' 'test_database_backup/test.py::test_database_backup_table[Disk('"'"'backup_disk_s3_plain'"'"', '"'"'test_table_backup'"'"')]' 'test_database_backup/test.py::test_database_backup_table[File('"'"'test_table_backup_file'"'"')]' 'test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_complex[complex_key_cache]' 'test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_complex[complex_key_direct]' 'test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_complex[complex_key_hashed]' 'test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_ranged[range_hashed]' 'test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_simple[cache]' 'test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_simple[direct]' 'test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_simple[flat]' 'test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_simple[hashed]' 'test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_complex[complex_key_cache]' 'test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_complex[complex_key_direct]' 'test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_complex[complex_key_hashed]' 'test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_ranged[range_hashed]' 'test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_simple[cache]' 'test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_simple[direct]' 'test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_simple[flat]' 'test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_simple[hashed]' 'test_dictionaries_redis/test.py::test_redis_dictionaries[DateTime_]' 'test_dictionaries_redis/test.py::test_redis_dictionaries[Date_]' 'test_dictionaries_redis/test.py::test_redis_dictionaries[Float32_]' 'test_dictionaries_redis/test.py::test_redis_dictionaries[Float64_]' 'test_dictionaries_redis/test.py::test_redis_dictionaries[Int16_]' 'test_dictionaries_redis/test.py::test_redis_dictionaries[Int32_]' 'test_dictionaries_redis/test.py::test_redis_dictionaries[Int64_]' 'test_dictionaries_redis/test.py::test_redis_dictionaries[Int8_]' 'test_dictionaries_redis/test.py::test_redis_dictionaries[String_]' -vvv -ss ENV CLICKHOUSE_LIBRARY_BRIDGE_BINARY_PATH /clickhouse-library-bridge ENV COMPOSE_HTTP_TIMEOUT 600 Stdout:net.ipv4.ip_local_port_range = 55000 65535 ENV DOCKER_MYSQL_PHP_CLIENT_TAG 88be89c1e3b6 ENV DOCKER_DOTNET_CLIENT_TAG 11de0b29a15d Source RedisSimple incompatible with layout complex_key_hashed ENV CLICKHOUSE_TESTS_CLIENT_BIN_PATH /clickhouse Source RedisSimple incompatible with layout complex_key_cache ENV DOCKER_MYSQL_JS_CLIENT_TAG 41ba7c2ec2a1 ENV PATH /spark-3.3.2-bin-hadoop3/bin:/opt/gdb/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin Command:[sysctl net.ipv4.ip_local_port_range='55000 65535'] Source RedisSimple incompatible with layout complex_key_direct ENV DOCKER_KERBERIZED_HADOOP_TAG latest Source RedisHash incompatible with layout flat Source RedisHash incompatible with layout hashed Running tests in /ClickHouse/tests/integration/test_config_substitutions/test.py ENV DOCKER_CHANNEL stable Source RedisHash incompatible with layout cache ENV DOCKER_CLIENT_TIMEOUT 300 Stdout:net.ipv4.ip_local_port_range = 55000 65535 ENV DOCKER_POSTGRESQL_JAVA_CLIENT_TAG a4eff5c7f4d6 Cluster start called. is_up=False ENV DOCKER_NGINX_DAV_TAG b55ac9cd7519 ENV DOCKER_MYSQL_GOLANG_CLIENT_TAG 9bec2a638e6e ENV PWD /ClickHouse/tests/integration ENV DOCKER_MYSQL_JAVA_CLIENT_TAG 766bff31cfe4 ENV CLICKHOUSE_ODBC_BRIDGE_BINARY_PATH /clickhouse-odbc-bridge Source RedisHash incompatible with layout direct ENV CLICKHOUSE_TESTS_BASE_CONFIG_DIR /clickhouse-config ENV TZ Etc/UTC Running tests in /ClickHouse/tests/integration/test_backup_restore_on_cluster/test.py ENV JAVA_PATH /usr/lib/jvm/java-11-openjdk-amd64/bin/java Cluster start called. is_up=False ENV DOCKER_BASE_TAG 8b2301119731 Command:[sysctl net.ipv4.ip_local_port_range='55000 65535'] ENV SPARK_HOME /spark-3.3.2-bin-hadoop3 ENV LC_CTYPE C.UTF-8 ENV INTEGRATION_TESTS_RUN_ID 0 Source RedisSimple incompatible with layout complex_key_hashed Source RedisSimple incompatible with layout complex_key_cache ENV WORKER_FREE_PORTS 30150 30151 30152 30153 30154 30155 30156 30157 30158 30159 30160 30161 30162 30163 30164 30165 30166 30167 30168 30169 30170 30171 30172 30173 30174 30175 30176 30177 30178 30179 30180 30181 30182 30183 30184 30185 30186 30187 30188 30189 30190 30191 30192 30193 30194 30195 30196 30197 30198 30199 ENV PYTEST_XDIST_TESTRUNUID 9b1b3f7145bb4931b6ac2b8301012c9f Stdout:net.ipv4.ip_local_port_range = 55000 65535 ENV PYTEST_XDIST_WORKER gw3 Source RedisSimple incompatible with layout complex_key_direct Stdout:net.ipv4.ip_local_port_range = 55000 65535 Source RedisHash incompatible with layout flat ENV PYTEST_XDIST_WORKER_COUNT 10 Source RedisHash incompatible with layout hashed Source RedisHash incompatible with layout cache Stdout:net.ipv4.ip_local_port_range = 55000 65535 ENV PYTEST_CURRENT_TEST test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_complex[complex_key_cache] (setup) Running tests in /ClickHouse/tests/integration/test_backup_restore/test.py Running tests in /ClickHouse/tests/integration/test_allow_feature_tier/test.py Source RedisHash incompatible with layout direct Cluster start called. is_up=False Cluster start called. is_up=False CLUSTER INIT base_config_dir:/clickhouse-config Source RedisSimple incompatible with layout complex_key_hashed Stdout:net.ipv4.ip_local_port_range = 55000 65535 Source RedisSimple incompatible with layout complex_key_cache Source RedisSimple incompatible with layout complex_key_direct Source RedisHash incompatible with layout flat Source RedisHash incompatible with layout hashed Source RedisHash incompatible with layout cache Running tests in /ClickHouse/tests/integration/test_cluster_all_replicas/test.py Cluster start called. is_up=False Source RedisHash incompatible with layout direct Source RedisSimple incompatible with layout complex_key_hashed Source RedisSimple incompatible with layout complex_key_cache Source RedisSimple incompatible with layout complex_key_direct Source RedisHash incompatible with layout flat Source RedisHash incompatible with layout hashed Source RedisHash incompatible with layout cache ENV DOCKER_KERBEROS_KDC_TAG 9391ecdee8d7 ENV CLICKHOUSE_TESTS_SERVER_BIN_PATH /clickhouse Source RedisHash incompatible with layout direct ENV MSAN_OPTIONS abort_on_error=1 poison_in_dtor=1 ENV JAVA_TOOL_OPTIONS -Djdk.attach.allowAttachSelf=true ENV TSAN_OPTIONS halt_on_error=1 abort_on_error=1 history_size=7 memory_limit_mb=46080 second_deadlock_stack=1 ENV HOSTNAME 1dd346380e80 clickhouse_start_command: clickhouse server --config-file=/etc/clickhouse-server/{main_config_file} --log-file=/var/log/clickhouse-server/clickhouse-server.log --errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log ENV SHLVL 0 ENV HOME /root ENV OLDPWD / ENV DOCKER_HELPER_TAG 5dc43a6382f0 Cluster name:mysql project_name:roottestdictionariesalllayoutsseparatesourcesmysql-gw3. Added instance name:node tag:8b2301119731 base_cmd:['docker', 'compose', '--env-file', '/ClickHouse/tests/integration/test_dictionaries_all_layouts_separate_sources/_instances-mysql-0-gw3/.env', '--project-name', 'roottestdictionariesalllayoutsseparatesourcesmysql-gw3', '--file', '/ClickHouse/tests/integration/test_dictionaries_all_layouts_separate_sources/_instances-mysql-0-gw3/node/docker-compose.yml', '--file', '/ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_mysql_8_0.yml'] docker_compose_yml_dir:/ClickHouse/tests/integration/helpers/../../../tests/integration/compose/ ENV PYTHONUNBUFFERED 1 ENV DOCKER_PYTHON_BOTTLE_TAG caad4729259e ENV UBSAN_OPTIONS print_stacktrace=1 Source RedisSimple incompatible with layout complex_key_hashed ENV PYTEST_ADDOPTS --dist=loadfile -n 10 -rfEps --run-id=0 --color=no --durations=0 test_aggregation_memory_efficient/test.py::test_remote test_allow_feature_tier/test.py::test_allow_feature_tier_in_general_settings test_allow_feature_tier/test.py::test_allow_feature_tier_in_mergetree_settings test_allow_feature_tier/test.py::test_allow_feature_tier_in_mergetree_settings_with_old_compatibility test_allow_feature_tier/test.py::test_allow_feature_tier_in_user test_allow_feature_tier/test.py::test_it_is_possible_to_enable_experimental_settings_in_default_profile test_analyzer_compatibility/test.py::test_two_new_versions test_async_load_databases/test.py::test_async_load_system_database test_async_load_databases/test.py::test_dependent_tables test_async_load_databases/test.py::test_dict_get_data test_async_load_databases/test.py::test_multiple_tables test_attach_table_normalizer/test.py::test_attach_substr test_attach_table_normalizer/test.py::test_attach_substr_restart test_azure_blob_storage_zero_copy_replication/test.py::test_zero_copy_replication test_backup_restore/test.py::test_attach_partition test_backup_restore/test.py::test_replace_partition test_backup_restore/test.py::test_restore 'test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[http-False]' 'test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[http-True]' 'test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[native-True]' test_backup_restore_on_cluster/test.py::test_backup_restore_on_single_replica test_backup_restore_on_cluster/test.py::test_different_tables_on_nodes test_backup_restore_on_cluster/test.py::test_empty_replicated_table test_backup_restore_on_cluster/test.py::test_file_deduplication test_backup_restore_on_cluster/test.py::test_get_error_from_other_host test_backup_restore_on_cluster/test.py::test_keeper_value_max_size test_backup_restore_on_cluster/test.py::test_mutation test_backup_restore_on_cluster/test.py::test_projection test_backup_restore_on_cluster/test.py::test_replicated_database test_backup_restore_on_cluster/test.py::test_replicated_database_async test_backup_restore_on_cluster/test.py::test_replicated_database_compare_parts 'test_backup_restore_on_cluster/test.py::test_replicated_database_with_special_macro_in_zk_path[database]' 'test_backup_restore_on_cluster/test.py::test_replicated_database_with_special_macro_in_zk_path[uuid]' test_backup_restore_on_cluster/test.py::test_replicated_table test_backup_restore_on_cluster/test.py::test_replicated_table_after_alters test_backup_restore_on_cluster/test.py::test_replicated_table_restored_into_bigger_cluster test_backup_restore_on_cluster/test.py::test_replicated_table_restored_into_smaller_cluster test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_def test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_insert test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_merge test_backup_restore_on_cluster/test.py::test_replicated_table_with_uuid_in_zkpath test_backup_restore_on_cluster/test.py::test_required_privileges test_backup_restore_on_cluster/test.py::test_shutdown_waits_for_backup test_backup_restore_on_cluster/test.py::test_system_functions test_backup_restore_on_cluster/test.py::test_system_users test_backup_restore_on_cluster/test.py::test_table_in_replicated_database_with_not_synced_def test_backup_restore_on_cluster/test.py::test_table_with_parts_in_queue_considered_non_empty test_backup_restore_on_cluster/test.py::test_tables_dependency test_cluster_all_replicas/test.py::test_cluster 'test_cluster_all_replicas/test.py::test_error_on_unavailable_replica[one_shard_three_nodes]' 'test_cluster_all_replicas/test.py::test_error_on_unavailable_replica[two_shards_three_nodes]' test_cluster_all_replicas/test.py::test_global_in 'test_cluster_all_replicas/test.py::test_skip_unavailable_replica[one_shard_three_nodes]' 'test_cluster_all_replicas/test.py::test_skip_unavailable_replica[two_shards_three_nodes]' test_composable_protocols/test.py::test_connections test_composable_protocols/test.py::test_http_proxy_1 test_composable_protocols/test.py::test_proxy_1 test_config_corresponding_root/test.py::test_work test_config_decryption/test.py::test_successful_decryption_xml test_config_decryption/test.py::test_successful_decryption_yaml test_config_hide_in_preprocessed/test.py::test_hide_in_preprocessed test_config_substitutions/test.py::test_allow_databases test_config_substitutions/test.py::test_config test_config_substitutions/test.py::test_config_from_env_overrides test_config_substitutions/test.py::test_config_merge_from_env_overrides test_config_substitutions/test.py::test_config_multiple_zk_substitutions test_config_substitutions/test.py::test_include_config 'test_database_backup/test.py::test_database_backup_database[Disk('"'"'backup_disk_local'"'"', '"'"'test_database_backup'"'"')]' 'test_database_backup/test.py::test_database_backup_database[Disk('"'"'backup_disk_object_storage_local_plain'"'"', '"'"'test_database_backup'"'"')]' 'test_database_backup/test.py::test_database_backup_database[Disk('"'"'backup_disk_s3_plain'"'"', '"'"'test_database_backup'"'"')]' 'test_database_backup/test.py::test_database_backup_database[File('"'"'test_database_backup_file'"'"')]' 'test_database_backup/test.py::test_database_backup_table[Disk('"'"'backup_disk_local'"'"', '"'"'test_table_backup'"'"')]' 'test_database_backup/test.py::test_database_backup_table[Disk('"'"'backup_disk_object_storage_local_plain'"'"', '"'"'test_table_backup'"'"')]' 'test_database_backup/test.py::test_database_backup_table[Disk('"'"'backup_disk_s3_plain'"'"', '"'"'test_table_backup'"'"')]' 'test_database_backup/test.py::test_database_backup_table[File('"'"'test_table_backup_file'"'"')]' 'test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_complex[complex_key_cache]' 'test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_complex[complex_key_direct]' 'test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_complex[complex_key_hashed]' 'test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_ranged[range_hashed]' 'test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_simple[cache]' 'test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_simple[direct]' 'test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_simple[flat]' 'test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_simple[hashed]' 'test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_complex[complex_key_cache]' 'test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_complex[complex_key_direct]' 'test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_complex[complex_key_hashed]' 'test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_ranged[range_hashed]' 'test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_simple[cache]' 'test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_simple[direct]' 'test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_simple[flat]' 'test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_simple[hashed]' 'test_dictionaries_redis/test.py::test_redis_dictionaries[DateTime_]' 'test_dictionaries_redis/test.py::test_redis_dictionaries[Date_]' 'test_dictionaries_redis/test.py::test_redis_dictionaries[Float32_]' 'test_dictionaries_redis/test.py::test_redis_dictionaries[Float64_]' 'test_dictionaries_redis/test.py::test_redis_dictionaries[Int16_]' 'test_dictionaries_redis/test.py::test_redis_dictionaries[Int32_]' 'test_dictionaries_redis/test.py::test_redis_dictionaries[Int64_]' 'test_dictionaries_redis/test.py::test_redis_dictionaries[Int8_]' 'test_dictionaries_redis/test.py::test_redis_dictionaries[String_]' -vvv -ss ENV CLICKHOUSE_LIBRARY_BRIDGE_BINARY_PATH /clickhouse-library-bridge ENV COMPOSE_HTTP_TIMEOUT 600 Running tests in /ClickHouse/tests/integration/test_dictionaries_all_layouts_separate_sources/test_mysql.py ENV DOCKER_MYSQL_PHP_CLIENT_TAG 88be89c1e3b6 ENV DOCKER_DOTNET_CLIENT_TAG 11de0b29a15d Source RedisSimple incompatible with layout complex_key_cache Cluster start called. is_up=False ENV CLICKHOUSE_TESTS_CLIENT_BIN_PATH /clickhouse ENV DOCKER_MYSQL_JS_CLIENT_TAG 41ba7c2ec2a1 ENV PATH /spark-3.3.2-bin-hadoop3/bin:/opt/gdb/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin ENV DOCKER_KERBERIZED_HADOOP_TAG latest ENV DOCKER_CHANNEL stable ENV DOCKER_CLIENT_TIMEOUT 300 ENV DOCKER_POSTGRESQL_JAVA_CLIENT_TAG a4eff5c7f4d6 Source RedisSimple incompatible with layout complex_key_direct ENV DOCKER_NGINX_DAV_TAG b55ac9cd7519 ENV DOCKER_MYSQL_GOLANG_CLIENT_TAG 9bec2a638e6e Source RedisHash incompatible with layout flat ENV PWD /ClickHouse/tests/integration Source RedisHash incompatible with layout hashed ENV DOCKER_MYSQL_JAVA_CLIENT_TAG 766bff31cfe4 Source RedisHash incompatible with layout cache ENV CLICKHOUSE_ODBC_BRIDGE_BINARY_PATH /clickhouse-odbc-bridge ENV CLICKHOUSE_TESTS_BASE_CONFIG_DIR /clickhouse-config ENV TZ Etc/UTC ENV JAVA_PATH /usr/lib/jvm/java-11-openjdk-amd64/bin/java ENV DOCKER_BASE_TAG 8b2301119731 ENV SPARK_HOME /spark-3.3.2-bin-hadoop3 ENV LC_CTYPE C.UTF-8 ENV INTEGRATION_TESTS_RUN_ID 0 ENV WORKER_FREE_PORTS 30250 30251 30252 30253 30254 30255 30256 30257 30258 30259 30260 30261 30262 30263 30264 30265 30266 30267 30268 30269 30270 30271 30272 30273 30274 30275 30276 30277 30278 30279 30280 30281 30282 30283 30284 30285 30286 30287 30288 30289 30290 30291 30292 30293 30294 30295 30296 30297 30298 30299 ENV PYTEST_XDIST_TESTRUNUID 9b1b3f7145bb4931b6ac2b8301012c9f ENV PYTEST_XDIST_WORKER gw5 ENV PYTEST_XDIST_WORKER_COUNT 10 ENV PYTEST_CURRENT_TEST test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_complex[complex_key_cache] (setup) Source RedisHash incompatible with layout direct CLUSTER INIT base_config_dir:/clickhouse-config clickhouse_start_command: clickhouse server --config-file=/etc/clickhouse-server/{main_config_file} --log-file=/var/log/clickhouse-server/clickhouse-server.log --errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log Cluster name:clickhouse_local project_name:roottestdictionariesalllayoutsseparatesourcesclickhouselocal-gw5. Added instance name:local_node tag:8b2301119731 base_cmd:['docker', 'compose', '--env-file', '/ClickHouse/tests/integration/test_dictionaries_all_layouts_separate_sources/_instances-clickhouse_local-0-gw5/.env', '--project-name', 'roottestdictionariesalllayoutsseparatesourcesclickhouselocal-gw5', '--file', '/ClickHouse/tests/integration/test_dictionaries_all_layouts_separate_sources/_instances-clickhouse_local-0-gw5/local_node/docker-compose.yml'] docker_compose_yml_dir:/ClickHouse/tests/integration/helpers/../../../tests/integration/compose/ Running tests in /ClickHouse/tests/integration/test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py Source RedisSimple incompatible with layout complex_key_hashed Cluster start called. is_up=False Source RedisSimple incompatible with layout complex_key_cache Source RedisSimple incompatible with layout complex_key_direct Source RedisHash incompatible with layout flat Source RedisHash incompatible with layout hashed Source RedisHash incompatible with layout cache Source RedisHash incompatible with layout direct Source RedisSimple incompatible with layout complex_key_hashed Source RedisSimple incompatible with layout complex_key_cache Source RedisSimple incompatible with layout complex_key_direct Source RedisHash incompatible with layout flat Source RedisHash incompatible with layout hashed Source RedisHash incompatible with layout cache Source RedisHash incompatible with layout direct Source RedisSimple incompatible with layout complex_key_hashed Source RedisSimple incompatible with layout complex_key_cache Source RedisSimple incompatible with layout complex_key_direct Source RedisHash incompatible with layout flat Source RedisHash incompatible with layout hashed Source RedisHash incompatible with layout cache Docker networks for project roottestallowfeaturetier-gw8 are NETWORK ID NAME DRIVER SCOPE Source RedisHash incompatible with layout direct Docker networks for project roottestbackuprestoreoncluster-gw2 are NETWORK ID NAME DRIVER SCOPE Source RedisSimple incompatible with layout complex_key_hashed Source RedisSimple incompatible with layout complex_key_cache Source RedisSimple incompatible with layout complex_key_direct Docker networks for project roottestdatabasebackup-gw1 are NETWORK ID NAME DRIVER SCOPE Source RedisHash incompatible with layout flat Source RedisHash incompatible with layout hashed Source RedisHash incompatible with layout cache Source RedisHash incompatible with layout direct Source RedisSimple incompatible with layout complex_key_hashed Source RedisSimple incompatible with layout complex_key_cache Source RedisSimple incompatible with layout complex_key_direct Source RedisHash incompatible with layout flat Source RedisHash incompatible with layout hashed Source RedisHash incompatible with layout cache Source RedisHash incompatible with layout direct Docker networks for project roottestbackuprestore-gw9 are NETWORK ID NAME DRIVER SCOPE Source RedisSimple incompatible with layout complex_key_hashed Source RedisSimple incompatible with layout complex_key_cache Source RedisSimple incompatible with layout complex_key_direct Source RedisHash incompatible with layout flat Source RedisHash incompatible with layout hashed Source RedisHash incompatible with layout cache Source RedisHash incompatible with layout direct Docker networks for project roottestclusterallreplicas-gw4 are NETWORK ID NAME DRIVER SCOPE Source RedisSimple incompatible with layout complex_key_hashed Source RedisSimple incompatible with layout complex_key_cache Source RedisSimple incompatible with layout complex_key_direct Source RedisHash incompatible with layout flat Source RedisHash incompatible with layout hashed Source RedisHash incompatible with layout cache Source RedisHash incompatible with layout direct Source RedisSimple incompatible with layout complex_key_hashed Source RedisSimple incompatible with layout complex_key_cache Source RedisSimple incompatible with layout complex_key_direct Source RedisHash incompatible with layout flat Source RedisHash incompatible with layout hashed Source RedisHash incompatible with layout cache Source RedisHash incompatible with layout direct Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisSimple_hashed_Float32_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisHash_complex_key_cache_UInt16_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisSimple_cache_Int8_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisSimple_flat_Int16_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisSimple_direct_String_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisHash_complex_key_cache_Int32_.xml Docker networks for project roottestasyncloaddatabases-gw7 are NETWORK ID NAME DRIVER SCOPE Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisSimple_cache_UInt64_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisSimple_flat_String_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisHash_complex_key_direct_UInt16_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisSimple_cache_UUID_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisSimple_hashed_Date_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisHash_complex_key_direct_String_.xml Docker networks for project roottestconfigsubstitutions-gw6 are NETWORK ID NAME DRIVER SCOPE Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisHash_complex_key_cache_Float32_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisHash_complex_key_cache_Int16_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisSimple_cache_Int32_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisHash_complex_key_cache_UInt64_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisHash_complex_key_direct_Int8_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisHash_complex_key_hashed_UInt8_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisHash_complex_key_hashed_Date_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisSimple_cache_Int16_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisSimple_cache_UInt8_.xml Docker networks for project roottestdictionariesalllayoutsseparatesourcesmysql-gw3 are NETWORK ID NAME DRIVER SCOPE Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisHash_complex_key_hashed_Float64_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisSimple_flat_Float64_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisHash_complex_key_hashed_UInt16_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisHash_complex_key_cache_Date_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisHash_complex_key_hashed_Int32_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisSimple_flat_Int64_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisSimple_hashed_Int8_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisSimple_direct_Int32_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisSimple_direct_UInt8_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisHash_complex_key_direct_UInt64_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisSimple_direct_UInt16_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisSimple_flat_Int32_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisHash_complex_key_hashed_Float32_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisHash_complex_key_hashed_Int64_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisSimple_cache_UInt32_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisSimple_flat_UInt16_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisSimple_flat_UInt64_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisHash_complex_key_direct_Float32_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisSimple_direct_UUID_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisSimple_direct_Int64_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisSimple_direct_Float64_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisHash_complex_key_cache_Int8_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisSimple_flat_Date_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisSimple_flat_UUID_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisSimple_hashed_UInt16_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisSimple_hashed_UInt32_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisHash_complex_key_hashed_Int16_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisSimple_hashed_DateTime_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisSimple_cache_UInt16_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisSimple_cache_Float32_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisSimple_hashed_UInt64_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisHash_complex_key_hashed_UUID_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisSimple_direct_Float32_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisHash_complex_key_direct_Int64_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisHash_complex_key_direct_UInt32_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisSimple_hashed_Float64_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisHash_complex_key_cache_String_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisHash_complex_key_hashed_UInt64_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisHash_complex_key_direct_DateTime_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisSimple_cache_DateTime_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisSimple_cache_String_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisSimple_direct_Date_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisSimple_cache_Float64_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisHash_complex_key_hashed_UInt32_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisSimple_direct_UInt32_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisSimple_hashed_String_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisSimple_flat_DateTime_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisSimple_direct_UInt64_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisHash_complex_key_cache_Int64_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisHash_complex_key_direct_Date_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisHash_complex_key_direct_Int16_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisHash_complex_key_direct_UUID_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisSimple_flat_Int8_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisHash_complex_key_direct_Float64_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisSimple_flat_Float32_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisSimple_flat_UInt32_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisHash_complex_key_cache_UInt32_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisHash_complex_key_hashed_String_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisHash_complex_key_cache_UUID_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisSimple_hashed_Int64_.xml Docker networks for project roottestdictionariesalllayoutsseparatesourcesclickhouselocal-gw5 are NETWORK ID NAME DRIVER SCOPE Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisSimple_hashed_UInt8_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisHash_complex_key_direct_UInt8_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisHash_complex_key_hashed_Int8_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisSimple_hashed_Int32_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisHash_complex_key_cache_Float64_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisSimple_hashed_UUID_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisSimple_direct_Int16_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisHash_complex_key_direct_Int32_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisSimple_cache_Date_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisSimple_cache_Int64_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisSimple_direct_Int8_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisHash_complex_key_cache_UInt8_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisHash_complex_key_hashed_DateTime_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisHash_complex_key_cache_DateTime_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisSimple_hashed_Int16_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisSimple_direct_DateTime_.xml Found dictionary /ClickHouse/tests/integration/test_dictionaries_redis/configs/dictionaries/RedisSimple_flat_UInt8_.xml clickhouse_start_command: clickhouse server --config-file=/etc/clickhouse-server/{main_config_file} --log-file=/var/log/clickhouse-server/clickhouse-server.log --errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log Docker containers for project roottestbackuprestoreoncluster-gw2 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Cluster name: project_name:roottestdictionariesredis-gw0. Added instance name:node tag:8b2301119731 base_cmd:['docker', 'compose', '--env-file', '/ClickHouse/tests/integration/test_dictionaries_redis/_instances-0-gw0/.env', '--project-name', 'roottestdictionariesredis-gw0', '--file', '/ClickHouse/tests/integration/test_dictionaries_redis/_instances-0-gw0/node/docker-compose.yml', '--file', '/ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_redis.yml'] docker_compose_yml_dir:/ClickHouse/tests/integration/helpers/../../../tests/integration/compose/ Running tests in /ClickHouse/tests/integration/test_dictionaries_redis/test.py Cluster start called. is_up=False Docker containers for project roottestdatabasebackup-gw1 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project roottestallowfeaturetier-gw8 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project roottestdictionariesalllayoutsseparatesourcesmysql-gw3 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project roottestbackuprestore-gw9 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project roottestdictionariesalllayoutsseparatesourcesclickhouselocal-gw5 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project roottestconfigsubstitutions-gw6 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project roottestclusterallreplicas-gw4 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker networks for project roottestdictionariesredis-gw0 are NETWORK ID NAME DRIVER SCOPE Docker volumes for project roottestdatabasebackup-gw1 are DRIVER VOLUME NAME Cleanup called Docker containers for project roottestasyncloaddatabases-gw7 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestbackuprestoreoncluster-gw2 are DRIVER VOLUME NAME Cleanup called Docker volumes for project roottestallowfeaturetier-gw8 are DRIVER VOLUME NAME Cleanup called Docker volumes for project roottestbackuprestore-gw9 are DRIVER VOLUME NAME Cleanup called Docker volumes for project roottestdictionariesalllayoutsseparatesourcesclickhouselocal-gw5 are DRIVER VOLUME NAME Cleanup called Docker volumes for project roottestdictionariesalllayoutsseparatesourcesmysql-gw3 are DRIVER VOLUME NAME Cleanup called Docker networks for project roottestdatabasebackup-gw1 are NETWORK ID NAME DRIVER SCOPE Docker volumes for project roottestconfigsubstitutions-gw6 are DRIVER VOLUME NAME Docker volumes for project roottestclusterallreplicas-gw4 are DRIVER VOLUME NAME Cleanup called Cleanup called Docker containers for project roottestdictionariesredis-gw0 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker networks for project roottestallowfeaturetier-gw8 are NETWORK ID NAME DRIVER SCOPE Docker volumes for project roottestasyncloaddatabases-gw7 are DRIVER VOLUME NAME Cleanup called Docker networks for project roottestdictionariesalllayoutsseparatesourcesclickhouselocal-gw5 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestdatabasebackup-gw1 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker networks for project roottestbackuprestore-gw9 are NETWORK ID NAME DRIVER SCOPE Docker networks for project roottestbackuprestoreoncluster-gw2 are NETWORK ID NAME DRIVER SCOPE Docker networks for project roottestclusterallreplicas-gw4 are NETWORK ID NAME DRIVER SCOPE Docker networks for project roottestdictionariesalllayoutsseparatesourcesmysql-gw3 are NETWORK ID NAME DRIVER SCOPE Docker networks for project roottestconfigsubstitutions-gw6 are NETWORK ID NAME DRIVER SCOPE Docker volumes for project roottestdictionariesredis-gw0 are DRIVER VOLUME NAME Cleanup called Docker containers for project roottestallowfeaturetier-gw8 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker networks for project roottestasyncloaddatabases-gw7 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestdictionariesalllayoutsseparatesourcesclickhouselocal-gw5 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project roottestclusterallreplicas-gw4 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project roottestbackuprestoreoncluster-gw2 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project roottestbackuprestore-gw9 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker containers for project roottestdictionariesalllayoutsseparatesourcesmysql-gw3 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker networks for project roottestdictionariesredis-gw0 are NETWORK ID NAME DRIVER SCOPE Docker volumes for project roottestdatabasebackup-gw1 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestdatabasebackup-gw1-.*-1$' --format '{{.ID}}:{{.Names}}'] Docker containers for project roottestconfigsubstitutions-gw6 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestallowfeaturetier-gw8 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestallowfeaturetier-gw8-.*-1$' --format '{{.ID}}:{{.Names}}'] Docker containers for project roottestasyncloaddatabases-gw7 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestbackuprestoreoncluster-gw2 are DRIVER VOLUME NAME Docker volumes for project roottestdictionariesalllayoutsseparatesourcesclickhouselocal-gw5 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestdictionariesalllayoutsseparatesourcesclickhouselocal-gw5-.*-1$' --format '{{.ID}}:{{.Names}}'] Command:[docker container list --all --filter name='^/roottestbackuprestoreoncluster-gw2-.*-1$' --format '{{.ID}}:{{.Names}}'] Docker volumes for project roottestclusterallreplicas-gw4 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestclusterallreplicas-gw4-.*-1$' --format '{{.ID}}:{{.Names}}'] Docker containers for project roottestdictionariesredis-gw0 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestconfigsubstitutions-gw6 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestconfigsubstitutions-gw6-.*-1$' --format '{{.ID}}:{{.Names}}'] Docker volumes for project roottestbackuprestore-gw9 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestbackuprestore-gw9-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} Unstopped containers: {} No running containers for project: roottestallowfeaturetier-gw8 Trying to prune unused networks... No running containers for project: roottestdatabasebackup-gw1 Trying to prune unused networks... Unstopped containers: {} No running containers for project: roottestdictionariesalllayoutsseparatesourcesclickhouselocal-gw5 Trying to prune unused networks... Unstopped containers: {} No running containers for project: roottestclusterallreplicas-gw4 Trying to prune unused networks... Docker volumes for project roottestasyncloaddatabases-gw7 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestasyncloaddatabases-gw7-.*-1$' --format '{{.ID}}:{{.Names}}'] Docker volumes for project roottestdictionariesredis-gw0 are DRIVER VOLUME NAME Unstopped containers: {} Command:[docker container list --all --filter name='^/roottestdictionariesredis-gw0-.*-1$' --format '{{.ID}}:{{.Names}}'] No running containers for project: roottestbackuprestoreoncluster-gw2 Trying to prune unused networks... Unstopped containers: {} No running containers for project: roottestbackuprestore-gw9 Trying to prune unused images... Trying to prune unused networks... Command:[docker image prune -f] Docker volumes for project roottestdictionariesalllayoutsseparatesourcesmysql-gw3 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestdictionariesalllayoutsseparatesourcesmysql-gw3-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} No running containers for project: roottestconfigsubstitutions-gw6 Trying to prune unused networks... Trying to prune unused images... Command:[docker image prune -f] Trying to prune unused images... Command:[docker image prune -f] Unstopped containers: {} Trying to prune unused images... No running containers for project: roottestasyncloaddatabases-gw7 Command:[docker image prune -f] Trying to prune unused networks... Unstopped containers: {} No running containers for project: roottestdictionariesredis-gw0 Trying to prune unused networks... Trying to prune unused images... Command:[docker image prune -f] Trying to prune unused images... Command:[docker image prune -f] Unstopped containers: {} No running containers for project: roottestdictionariesalllayoutsseparatesourcesmysql-gw3 Trying to prune unused networks... Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Trying to prune unused images... Command:[docker image prune -f] Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Trying to prune unused images... Command:[docker image prune -f] Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Trying to prune unused images... Stdout:Total reclaimed space: 0B Command:[docker image prune -f] Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Stderr:Error response from daemon: a prune operation is already running Exitcode:1 Trying to prune unused volumes... Command:[docker volume ls | wc -l] Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Stdout:1 Volumes pruned: 1 Setup directory for instance: instance Create directory for configuration generated in this helper Create directory for common tests configuration Copy common configuration from helpers Generate and write macros file Copy custom test config files ['/ClickHouse/tests/integration/test_allow_feature_tier/configs/allow_feature_tier.xml'] to /ClickHouse/tests/integration/test_allow_feature_tier/_instances-0-gw8/instance/configs/config.d Setup database dir /ClickHouse/tests/integration/test_allow_feature_tier/_instances-0-gw8/instance/database Setup logs dir /ClickHouse/tests/integration/test_allow_feature_tier/_instances-0-gw8/instance/logs Entrypoint cmd: bash -c "trap 'pkill tail' INT TERM; clickhouse server --config-file=/etc/clickhouse-server/config.xml --log-file=/var/log/clickhouse-server/clickhouse-server.log --errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log --daemon -- ; coproc tail -f /dev/null; wait $$!" Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw'} stored in /ClickHouse/tests/integration/test_allow_feature_tier/_instances-0-gw8/.env Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found Trying to prune unused images... Command:[docker image prune -f] Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found Stdout:1 Volumes pruned: 1 Setup directory for instance: local_node Create directory for configuration generated in this helper Create directory for common tests configuration Copy common configuration from helpers Stdout:1 Generate and write macros file Stderr:Error response from daemon: a prune operation is already running Volumes pruned: 1 Setup directory for instance: instance Copy custom test config files ['/ClickHouse/tests/integration/test_dictionaries_all_layouts_separate_sources/configs/disable_ssl_verification.xml'] to /ClickHouse/tests/integration/test_dictionaries_all_layouts_separate_sources/_instances-clickhouse_local-0-gw5/local_node/configs/config.d Exitcode:1 Trying to prune unused volumes... Command:[docker volume ls | wc -l] Create directory for configuration generated in this helper Create directory for common tests configuration Stdout:Total reclaimed space: 0B Copy common configuration from helpers Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Generate and write macros file Copy custom test config files ['/ClickHouse/tests/integration/test_database_backup/configs/backups.xml'] to /ClickHouse/tests/integration/test_database_backup/_instances-0-gw1/instance/configs/config.d Setup database dir /ClickHouse/tests/integration/test_dictionaries_all_layouts_separate_sources/_instances-clickhouse_local-0-gw5/local_node/database Setup logs dir /ClickHouse/tests/integration/test_dictionaries_all_layouts_separate_sources/_instances-clickhouse_local-0-gw5/local_node/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw'} stored in /ClickHouse/tests/integration/test_dictionaries_all_layouts_separate_sources/_instances-clickhouse_local-0-gw5/.env Setup database dir /ClickHouse/tests/integration/test_database_backup/_instances-0-gw1/instance/database Setup logs dir /ClickHouse/tests/integration/test_database_backup/_instances-0-gw1/instance/logs Entrypoint cmd: bash -c "trap 'pkill tail' INT TERM; clickhouse server --config-file=/etc/clickhouse-server/config.xml --log-file=/var/log/clickhouse-server/clickhouse-server.log --errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log --daemon -- ; coproc tail -f /dev/null; wait $$!" Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw', 'MINIO_CERTS_DIR': '/ClickHouse/tests/integration/test_database_backup/_instances-0-gw1/minio/certs', 'MINIO_DATA_DIR': '/ClickHouse/tests/integration/test_database_backup/_instances-0-gw1/minio/data', 'MINIO_PORT': '9001', 'SSL_CERT_FILE': '/ClickHouse/tests/integration/test_database_backup/_instances-0-gw1/minio/certs/public.crt', 'RESOLVER_LOGS': '/ClickHouse/tests/integration/test_database_backup/_instances-0-gw1/resolver', 'RESOLVER_LOGS_FS': 'bind'} stored in /ClickHouse/tests/integration/test_database_backup/_instances-0-gw1/.env No config file found Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found Stdout:1 http://localhost:None "GET /version HTTP/1.1" 200 826 Volumes pruned: 1 Setup directory for instance: node1 Command:[docker compose --env-file /ClickHouse/tests/integration/test_allow_feature_tier/_instances-0-gw8/.env --project-name roottestallowfeaturetier-gw8 --file /ClickHouse/tests/integration/test_allow_feature_tier/_instances-0-gw8/instance/docker-compose.yml pull] Create directory for configuration generated in this helper Create directory for common tests configuration Copy common configuration from helpers Generate and write macros file Copy custom test config files ['/ClickHouse/tests/integration/test_cluster_all_replicas/configs/remote_servers.xml'] to /ClickHouse/tests/integration/test_cluster_all_replicas/_instances-0-gw4/node1/configs/config.d Setup database dir /ClickHouse/tests/integration/test_cluster_all_replicas/_instances-0-gw4/node1/database Setup logs dir /ClickHouse/tests/integration/test_cluster_all_replicas/_instances-0-gw4/node1/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Setup directory for instance: node2 Create directory for configuration generated in this helper Create directory for common tests configuration Copy common configuration from helpers Stdout:1 Stdout:1 Volumes pruned: 1 Setup directory for instance: node1 Generate and write macros file Volumes pruned: 1 Setup directory for instance: node Copy custom test config files ['/ClickHouse/tests/integration/test_cluster_all_replicas/configs/remote_servers.xml'] to /ClickHouse/tests/integration/test_cluster_all_replicas/_instances-0-gw4/node2/configs/config.d Setup database dir /ClickHouse/tests/integration/test_cluster_all_replicas/_instances-0-gw4/node2/database Create directory for configuration generated in this helper Setup logs dir /ClickHouse/tests/integration/test_cluster_all_replicas/_instances-0-gw4/node2/logs Create directory for common tests configuration Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw'} stored in /ClickHouse/tests/integration/test_cluster_all_replicas/_instances-0-gw4/.env Copy common configuration from helpers Create directory for configuration generated in this helper Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] Create directory for common tests configuration No config file found Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] Generate and write macros file Copy common configuration from helpers No config file found Copy custom test config files ['/ClickHouse/tests/integration/test_backup_restore_on_cluster/configs/cluster.xml', '/ClickHouse/tests/integration/test_backup_restore_on_cluster/configs/cluster3.xml', '/ClickHouse/tests/integration/test_backup_restore_on_cluster/configs/replicated_access_storage.xml', '/ClickHouse/tests/integration/test_backup_restore_on_cluster/configs/replicated_user_defined_sql_objects.xml', '/ClickHouse/tests/integration/test_backup_restore_on_cluster/configs/backups_disk.xml', '/ClickHouse/tests/integration/test_backup_restore_on_cluster/configs/lesser_timeouts.xml'] to /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw2/node1/configs/config.d Generate and write macros file Copy custom test config files [] to /ClickHouse/tests/integration/test_backup_restore/_instances-0-gw9/node/configs/config.d Setup database dir /ClickHouse/tests/integration/test_backup_restore/_instances-0-gw9/node/database Setup logs dir /ClickHouse/tests/integration/test_backup_restore/_instances-0-gw9/node/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw'} stored in /ClickHouse/tests/integration/test_backup_restore/_instances-0-gw9/.env Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found http://localhost:None "GET /version HTTP/1.1" 200 826 Command:[docker compose --env-file /ClickHouse/tests/integration/test_database_backup/_instances-0-gw1/.env --project-name roottestdatabasebackup-gw1 --file /ClickHouse/tests/integration/test_database_backup/_instances-0-gw1/instance/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_minio.yml pull] Setup database dir /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw2/node1/database Setup logs dir /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw2/node1/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] external_dir_abs_path=/ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw2/backups http://localhost:None "GET /version HTTP/1.1" 200 826 Setup directory for instance: node2 Create directory for configuration generated in this helper Command:[docker compose --env-file /ClickHouse/tests/integration/test_dictionaries_all_layouts_separate_sources/_instances-clickhouse_local-0-gw5/.env --project-name roottestdictionariesalllayoutsseparatesourcesclickhouselocal-gw5 --file /ClickHouse/tests/integration/test_dictionaries_all_layouts_separate_sources/_instances-clickhouse_local-0-gw5/local_node/docker-compose.yml pull] Create directory for common tests configuration Copy common configuration from helpers Generate and write macros file Copy custom test config files ['/ClickHouse/tests/integration/test_backup_restore_on_cluster/configs/cluster.xml', '/ClickHouse/tests/integration/test_backup_restore_on_cluster/configs/cluster3.xml', '/ClickHouse/tests/integration/test_backup_restore_on_cluster/configs/replicated_access_storage.xml', '/ClickHouse/tests/integration/test_backup_restore_on_cluster/configs/replicated_user_defined_sql_objects.xml', '/ClickHouse/tests/integration/test_backup_restore_on_cluster/configs/backups_disk.xml', '/ClickHouse/tests/integration/test_backup_restore_on_cluster/configs/lesser_timeouts.xml'] to /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw2/node2/configs/config.d Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Setup database dir /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw2/node2/database Setup logs dir /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw2/node2/logs Stdout:Total reclaimed space: 0B Entrypoint cmd: bash -c "trap 'pkill tail' INT TERM; clickhouse server --config-file=/etc/clickhouse-server/config.xml --log-file=/var/log/clickhouse-server/clickhouse-server.log --errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log --daemon -- ; coproc tail -f /dev/null; wait $$!" Images pruned Trying to prune unused volumes... external_dir_abs_path=/ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw2/backups Command:[docker volume ls | wc -l] Setup directory for instance: node3 Create directory for configuration generated in this helper Create directory for common tests configuration Copy common configuration from helpers Generate and write macros file Copy custom test config files ['/ClickHouse/tests/integration/test_backup_restore_on_cluster/configs/cluster.xml', '/ClickHouse/tests/integration/test_backup_restore_on_cluster/configs/cluster3.xml', '/ClickHouse/tests/integration/test_backup_restore_on_cluster/configs/replicated_access_storage.xml', '/ClickHouse/tests/integration/test_backup_restore_on_cluster/configs/replicated_user_defined_sql_objects.xml', '/ClickHouse/tests/integration/test_backup_restore_on_cluster/configs/backups_disk.xml', '/ClickHouse/tests/integration/test_backup_restore_on_cluster/configs/lesser_timeouts.xml'] to /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw2/node3/configs/config.d Setup database dir /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw2/node3/database Setup logs dir /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw2/node3/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] external_dir_abs_path=/ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw2/backups Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw', 'keeper_binary': '/clickhouse', 'keeper_cmd_prefix': 'clickhouse keeper', 'image': 'altinityinfra/integration-test:8b2301119731', 'user': '0', 'keeper_fs': 'bind', 'keeper_logs_dir1': '/ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw2/keeper1/log', 'keeper_config_dir1': '/ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw2/keeper1/config', 'keeper_db_dir1': '/ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw2/keeper1/coordination', 'keeper_logs_dir2': '/ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw2/keeper2/log', 'keeper_config_dir2': '/ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw2/keeper2/config', 'keeper_db_dir2': '/ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw2/keeper2/coordination', 'keeper_logs_dir3': '/ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw2/keeper3/log', 'keeper_config_dir3': '/ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw2/keeper3/config', 'keeper_db_dir3': '/ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw2/keeper3/coordination'} stored in /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw2/.env Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found Stdout:1 Volumes pruned: 1 Setup directory for instance: node1 Create directory for configuration generated in this helper Create directory for common tests configuration Copy common configuration from helpers Generate and write macros file Copy custom test config files ['/ClickHouse/tests/integration/test_async_load_databases/configs/config.xml'] to /ClickHouse/tests/integration/test_async_load_databases/_instances-0-gw7/node1/configs/config.d http://localhost:None "GET /version HTTP/1.1" 200 826 Command:[docker compose --env-file /ClickHouse/tests/integration/test_cluster_all_replicas/_instances-0-gw4/.env --project-name roottestclusterallreplicas-gw4 --file /ClickHouse/tests/integration/test_cluster_all_replicas/_instances-0-gw4/node1/docker-compose.yml --file /ClickHouse/tests/integration/test_cluster_all_replicas/_instances-0-gw4/node2/docker-compose.yml pull] Setup database dir /ClickHouse/tests/integration/test_async_load_databases/_instances-0-gw7/node1/database Setup logs dir /ClickHouse/tests/integration/test_async_load_databases/_instances-0-gw7/node1/logs Entrypoint cmd: bash -c "trap 'pkill tail' INT TERM; clickhouse server --config-file=/etc/clickhouse-server/config.xml --log-file=/var/log/clickhouse-server/clickhouse-server.log --errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log --daemon -- ; coproc tail -f /dev/null; wait $$!" Stdout:1 Setup directory for instance: node2 Volumes pruned: 1 Setup directory for instance: node1 Create directory for configuration generated in this helper Create directory for configuration generated in this helper Create directory for common tests configuration Copy common configuration from helpers Create directory for common tests configuration Copy common configuration from helpers Generate and write macros file Copy custom test config files [] to /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw6/node1/configs/config.d Generate and write macros file Copy custom test config files ['/ClickHouse/tests/integration/test_async_load_databases/configs/async_load_system_database.xml'] to /ClickHouse/tests/integration/test_async_load_databases/_instances-0-gw7/node2/configs/config.d http://localhost:None "GET /version HTTP/1.1" 200 826 Command:[docker compose --env-file /ClickHouse/tests/integration/test_backup_restore/_instances-0-gw9/.env --project-name roottestbackuprestore-gw9 --file /ClickHouse/tests/integration/test_backup_restore/_instances-0-gw9/node/docker-compose.yml pull] Setup database dir /ClickHouse/tests/integration/test_async_load_databases/_instances-0-gw7/node2/database Setup database dir /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw6/node1/database Setup logs dir /ClickHouse/tests/integration/test_async_load_databases/_instances-0-gw7/node2/logs Setup logs dir /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw6/node1/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Entrypoint cmd: bash -c "trap 'pkill tail' INT TERM; clickhouse server --config-file=/etc/clickhouse-server/config.xml --log-file=/var/log/clickhouse-server/clickhouse-server.log --errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log --daemon -- ; coproc tail -f /dev/null; wait $$!" Setup directory for instance: node2 Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw'} stored in /ClickHouse/tests/integration/test_async_load_databases/_instances-0-gw7/.env Create directory for configuration generated in this helper Create directory for common tests configuration Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] Copy common configuration from helpers No config file found Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] Generate and write macros file No config file found Copy custom test config files [] to /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw6/node2/configs/config.d Setup database dir /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw6/node2/database Setup logs dir /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw6/node2/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] http://localhost:None "GET /version HTTP/1.1" 200 826 Setup directory for instance: node3 Create directory for configuration generated in this helper Create directory for common tests configuration Copy common configuration from helpers Command:[docker compose --env-file /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw2/.env --project-name roottestbackuprestoreoncluster-gw2 --file /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw2/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw2/node2/docker-compose.yml --file /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw2/node3/docker-compose.yml pull] Generate and write macros file Copy custom test config files ['/ClickHouse/tests/integration/test_config_substitutions/configs/config_zk_include_test.xml'] to /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw6/node3/configs/config.d Setup database dir /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw6/node3/database Setup logs dir /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw6/node3/logs Entrypoint cmd: bash -c "trap 'pkill tail' INT TERM; clickhouse server --config-file=/etc/clickhouse-server/config.xml --log-file=/var/log/clickhouse-server/clickhouse-server.log --errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log --daemon -- ; coproc tail -f /dev/null; wait $$!" Setup directory for instance: node4 Create directory for configuration generated in this helper Create directory for common tests configuration Stdout:1 Copy common configuration from helpers Volumes pruned: 1 Setup directory for instance: node Generate and write macros file Copy custom test config files ['/ClickHouse/tests/integration/test_config_substitutions/configs/include_from_source.xml'] to /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw6/node4/configs/config.d Create directory for configuration generated in this helper Stdout:1 Create directory for common tests configuration Volumes pruned: 1 Setup directory for instance: node Copy common configuration from helpers Create directory for configuration generated in this helper Create directory for common tests configuration Copy common configuration from helpers Generate and write macros file Copy custom test config files [] to /ClickHouse/tests/integration/test_dictionaries_redis/_instances-0-gw0/node/configs/config.d Generate and write macros file Copy custom test config files ['/ClickHouse/tests/integration/test_dictionaries_all_layouts_separate_sources/configs/disable_ssl_verification.xml'] to /ClickHouse/tests/integration/test_dictionaries_all_layouts_separate_sources/_instances-mysql-0-gw3/node/configs/config.d Setup database dir /ClickHouse/tests/integration/test_dictionaries_all_layouts_separate_sources/_instances-mysql-0-gw3/node/database Setup logs dir /ClickHouse/tests/integration/test_dictionaries_all_layouts_separate_sources/_instances-mysql-0-gw3/node/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw', 'MYSQL8_HOST': 'mysql80', 'MYSQL8_PORT': '3306', 'MYSQL8_ROOT_HOST': '%', 'MYSQL8_LOGS': '/ClickHouse/tests/integration/test_dictionaries_all_layouts_separate_sources/_instances-mysql-0-gw3/mysql8/logs', 'MYSQL8_LOGS_FS': 'bind', 'MYSQL8_DOCKER_USER': '0'} stored in /ClickHouse/tests/integration/test_dictionaries_all_layouts_separate_sources/_instances-mysql-0-gw3/.env Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found Setup database dir /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw6/node4/database Setup logs dir /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw6/node4/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Setup directory for instance: node5 Create directory for configuration generated in this helper Create directory for common tests configuration Copy common configuration from helpers Generate and write macros file Copy custom test config files [] to /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw6/node5/configs/config.d Setup database dir /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw6/node5/database Setup logs dir /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw6/node5/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Setup directory for instance: node6 Create directory for configuration generated in this helper Create directory for common tests configuration http://localhost:None "GET /version HTTP/1.1" 200 826 Copy common configuration from helpers Setup database dir /ClickHouse/tests/integration/test_dictionaries_redis/_instances-0-gw0/node/database Command:[docker compose --env-file /ClickHouse/tests/integration/test_async_load_databases/_instances-0-gw7/.env --project-name roottestasyncloaddatabases-gw7 --file /ClickHouse/tests/integration/test_async_load_databases/_instances-0-gw7/node1/docker-compose.yml --file /ClickHouse/tests/integration/test_async_load_databases/_instances-0-gw7/node2/docker-compose.yml pull] Generate and write macros file Setup logs dir /ClickHouse/tests/integration/test_dictionaries_redis/_instances-0-gw0/node/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Copy custom test config files ['/ClickHouse/tests/integration/test_config_substitutions/configs/include_from_source.xml'] to /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw6/node6/configs/config.d Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw', 'REDIS_HOST': 'redis1', 'REDIS_EXTERNAL_PORT': '30000', 'REDIS_INTERNAL_PORT': '6379'} stored in /ClickHouse/tests/integration/test_dictionaries_redis/_instances-0-gw0/.env http://localhost:None "GET /version HTTP/1.1" 200 826 Setup database dir /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw6/node6/database Setup logs dir /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw6/node6/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Command:[docker compose --env-file /ClickHouse/tests/integration/test_dictionaries_all_layouts_separate_sources/_instances-mysql-0-gw3/.env --project-name roottestdictionariesalllayoutsseparatesourcesmysql-gw3 --file /ClickHouse/tests/integration/test_dictionaries_all_layouts_separate_sources/_instances-mysql-0-gw3/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_mysql_8_0.yml pull] Setup directory for instance: node7 Create directory for configuration generated in this helper Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] Create directory for common tests configuration No config file found Copy common configuration from helpers Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found Generate and write macros file Copy custom test config files ['/ClickHouse/tests/integration/test_config_substitutions/configs/000-server_overrides.xml', '/ClickHouse/tests/integration/test_config_substitutions/configs/010-server_with_env_subst.xml'] to /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw6/node7/configs/config.d Setup database dir /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw6/node7/database Setup logs dir /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw6/node7/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw', 'MAX_QUERY_SIZE': '121212', 'keeper_binary': '/clickhouse', 'keeper_cmd_prefix': 'clickhouse keeper', 'image': 'altinityinfra/integration-test:8b2301119731', 'user': '0', 'keeper_fs': 'bind', 'keeper_logs_dir1': '/ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw6/keeper1/log', 'keeper_config_dir1': '/ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw6/keeper1/config', 'keeper_db_dir1': '/ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw6/keeper1/coordination', 'keeper_logs_dir2': '/ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw6/keeper2/log', 'keeper_config_dir2': '/ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw6/keeper2/config', 'keeper_db_dir2': '/ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw6/keeper2/coordination', 'keeper_logs_dir3': '/ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw6/keeper3/log', 'keeper_config_dir3': '/ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw6/keeper3/config', 'keeper_db_dir3': '/ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw6/keeper3/coordination', 'INCLUDE_FROM_ENV': '/etc/clickhouse-server/config.d/include_from_source.xml', 'MAX_THREADS': '2'} stored in /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw6/node7/.env Setup directory for instance: node8 Create directory for configuration generated in this helper Create directory for common tests configuration Copy common configuration from helpers Generate and write macros file Copy custom test config files ['/ClickHouse/tests/integration/test_config_substitutions/configs/include_from_source.yml'] to /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw6/node8/configs/config.d Setup database dir /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw6/node8/database Setup logs dir /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw6/node8/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw', 'MAX_QUERY_SIZE': '55555', 'keeper_binary': '/clickhouse', 'keeper_cmd_prefix': 'clickhouse keeper', 'image': 'altinityinfra/integration-test:8b2301119731', 'user': '0', 'keeper_fs': 'bind', 'keeper_logs_dir1': '/ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw6/keeper1/log', 'keeper_config_dir1': '/ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw6/keeper1/config', 'keeper_db_dir1': '/ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw6/keeper1/coordination', 'keeper_logs_dir2': '/ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw6/keeper2/log', 'keeper_config_dir2': '/ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw6/keeper2/config', 'keeper_db_dir2': '/ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw6/keeper2/coordination', 'keeper_logs_dir3': '/ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw6/keeper3/log', 'keeper_config_dir3': '/ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw6/keeper3/config', 'keeper_db_dir3': '/ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw6/keeper3/coordination', 'INCLUDE_FROM_ENV': '/etc/clickhouse-server/config.d/include_from_source.xml'} stored in /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw6/.env Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found http://localhost:None "GET /version HTTP/1.1" 200 826 Command:[docker compose --env-file /ClickHouse/tests/integration/test_dictionaries_redis/_instances-0-gw0/.env --project-name roottestdictionariesredis-gw0 --file /ClickHouse/tests/integration/test_dictionaries_redis/_instances-0-gw0/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_redis.yml pull] http://localhost:None "GET /version HTTP/1.1" 200 826 Command:[docker compose --env-file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw6/.env --project-name roottestconfigsubstitutions-gw6 --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw6/node1/docker-compose.yml --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw6/node2/docker-compose.yml --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw6/node3/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw6/node4/docker-compose.yml --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw6/node5/docker-compose.yml --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw6/node6/docker-compose.yml --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw6/node7/docker-compose.yml --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw6/node8/docker-compose.yml pull] Stderr: node Pulling Stderr: node Pulled ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_backup_restore/_instances-0-gw9/.env --project-name roottestbackuprestore-gw9 --file /ClickHouse/tests/integration/test_backup_restore/_instances-0-gw9/node/docker-compose.yml up -d --no-recreate') Command:[docker compose --env-file /ClickHouse/tests/integration/test_backup_restore/_instances-0-gw9/.env --project-name roottestbackuprestore-gw9 --file /ClickHouse/tests/integration/test_backup_restore/_instances-0-gw9/node/docker-compose.yml up -d --no-recreate] Stderr: node1 Skipped - Image is already being pulled by node2 Stderr: node2 Pulling Stderr: node2 Pulled ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_async_load_databases/_instances-0-gw7/.env --project-name roottestasyncloaddatabases-gw7 --file /ClickHouse/tests/integration/test_async_load_databases/_instances-0-gw7/node1/docker-compose.yml --file /ClickHouse/tests/integration/test_async_load_databases/_instances-0-gw7/node2/docker-compose.yml up -d --no-recreate') Command:[docker compose --env-file /ClickHouse/tests/integration/test_async_load_databases/_instances-0-gw7/.env --project-name roottestasyncloaddatabases-gw7 --file /ClickHouse/tests/integration/test_async_load_databases/_instances-0-gw7/node1/docker-compose.yml --file /ClickHouse/tests/integration/test_async_load_databases/_instances-0-gw7/node2/docker-compose.yml up -d --no-recreate] Stderr: local_node Pulling Stderr: local_node Pulled ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_dictionaries_all_layouts_separate_sources/_instances-clickhouse_local-0-gw5/.env --project-name roottestdictionariesalllayoutsseparatesourcesclickhouselocal-gw5 --file /ClickHouse/tests/integration/test_dictionaries_all_layouts_separate_sources/_instances-clickhouse_local-0-gw5/local_node/docker-compose.yml up -d --no-recreate') Command:[docker compose --env-file /ClickHouse/tests/integration/test_dictionaries_all_layouts_separate_sources/_instances-clickhouse_local-0-gw5/.env --project-name roottestdictionariesalllayoutsseparatesourcesclickhouselocal-gw5 --file /ClickHouse/tests/integration/test_dictionaries_all_layouts_separate_sources/_instances-clickhouse_local-0-gw5/local_node/docker-compose.yml up -d --no-recreate] Stderr: instance Pulling Stderr: instance Pulled ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_allow_feature_tier/_instances-0-gw8/.env --project-name roottestallowfeaturetier-gw8 --file /ClickHouse/tests/integration/test_allow_feature_tier/_instances-0-gw8/instance/docker-compose.yml up -d --no-recreate') Command:[docker compose --env-file /ClickHouse/tests/integration/test_allow_feature_tier/_instances-0-gw8/.env --project-name roottestallowfeaturetier-gw8 --file /ClickHouse/tests/integration/test_allow_feature_tier/_instances-0-gw8/instance/docker-compose.yml up -d --no-recreate] Stderr: node2 Skipped - Image is already being pulled by zoo3 Stderr: node3 Skipped - Image is already being pulled by zoo3 Stderr: node1 Skipped - Image is already being pulled by zoo3 Stderr: zoo1 Skipped - Image is already being pulled by zoo3 Stderr: zoo2 Skipped - Image is already being pulled by zoo3 Stderr: zoo3 Pulling Stderr: zoo3 Pulled Setup ZooKeeper Creating internal ZooKeeper dirs: ['/ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw2/keeper1/log', '/ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw2/keeper1/config', '/ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw2/keeper1/coordination', '/ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw2/keeper2/log', '/ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw2/keeper2/config', '/ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw2/keeper2/coordination', '/ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw2/keeper3/log', '/ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw2/keeper3/config', '/ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw2/keeper3/coordination'] Command:[docker compose --project-name roottestbackuprestoreoncluster-gw2 --env-file /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw2/.env --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --verbose up -d] Stderr: node1 Skipped - Image is already being pulled by node2 Stderr: node2 Pulling Stderr: node2 Pulled ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_cluster_all_replicas/_instances-0-gw4/.env --project-name roottestclusterallreplicas-gw4 --file /ClickHouse/tests/integration/test_cluster_all_replicas/_instances-0-gw4/node1/docker-compose.yml --file /ClickHouse/tests/integration/test_cluster_all_replicas/_instances-0-gw4/node2/docker-compose.yml up -d --no-recreate') Command:[docker compose --env-file /ClickHouse/tests/integration/test_cluster_all_replicas/_instances-0-gw4/.env --project-name roottestclusterallreplicas-gw4 --file /ClickHouse/tests/integration/test_cluster_all_replicas/_instances-0-gw4/node1/docker-compose.yml --file /ClickHouse/tests/integration/test_cluster_all_replicas/_instances-0-gw4/node2/docker-compose.yml up -d --no-recreate] Stderr: node1 Skipped - Image is already being pulled by node6 Stderr: node5 Skipped - Image is already being pulled by node6 Stderr: zoo1 Skipped - Image is already being pulled by node6 Stderr: node7 Skipped - Image is already being pulled by node6 Stderr: node8 Skipped - Image is already being pulled by node6 Stderr: zoo2 Skipped - Image is already being pulled by node6 Stderr: zoo3 Skipped - Image is already being pulled by node6 Stderr: node2 Skipped - Image is already being pulled by node6 Stderr: node3 Skipped - Image is already being pulled by node6 Stderr: node4 Skipped - Image is already being pulled by node6 Stderr: node6 Pulling Stderr: node6 Pulled Setup ZooKeeper Creating internal ZooKeeper dirs: ['/ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw6/keeper1/log', '/ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw6/keeper1/config', '/ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw6/keeper1/coordination', '/ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw6/keeper2/log', '/ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw6/keeper2/config', '/ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw6/keeper2/coordination', '/ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw6/keeper3/log', '/ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw6/keeper3/config', '/ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw6/keeper3/coordination'] Command:[docker compose --project-name roottestconfigsubstitutions-gw6 --env-file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw6/.env --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --verbose up -d] Stderr: Network roottestbackuprestore-gw9_default Creating Stderr: Network roottestbackuprestore-gw9_default Created Stderr: Container roottestbackuprestore-gw9-node-1 Creating Stderr: Container roottestbackuprestore-gw9-node-1 Created Stderr: Container roottestbackuprestore-gw9-node-1 Starting Stderr: Container roottestbackuprestore-gw9-node-1 Started ClickHouse instance created get_instance_ip instance_name=node http://localhost:None "GET /v1.46/containers/roottestbackuprestore-gw9-node-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node http://localhost:None "GET /v1.46/containers/roottestbackuprestore-gw9-node-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node, ip: 172.16.1.2... http://localhost:None "GET /v1.46/containers/roottestbackuprestore-gw9-node-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/ed1c3dad61920fe13a676bda7d47ba57d1baa9a4198dbe901f7031f9d256afa3/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/ed1c3dad61920fe13a676bda7d47ba57d1baa9a4198dbe901f7031f9d256afa3/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/ed1c3dad61920fe13a676bda7d47ba57d1baa9a4198dbe901f7031f9d256afa3/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/ed1c3dad61920fe13a676bda7d47ba57d1baa9a4198dbe901f7031f9d256afa3/json HTTP/1.1" 200 None Stderr: Network roottestasyncloaddatabases-gw7_default Creating Stderr: Network roottestasyncloaddatabases-gw7_default Created Stderr: Container roottestasyncloaddatabases-gw7-node1-1 Creating Stderr: Container roottestasyncloaddatabases-gw7-node2-1 Creating Stderr: Container roottestasyncloaddatabases-gw7-node2-1 Created Stderr: Container roottestasyncloaddatabases-gw7-node1-1 Created Stderr: Container roottestasyncloaddatabases-gw7-node1-1 Starting Stderr: Container roottestasyncloaddatabases-gw7-node2-1 Starting Stderr: Container roottestasyncloaddatabases-gw7-node1-1 Started Stderr: Container roottestasyncloaddatabases-gw7-node2-1 Started ClickHouse instance created get_instance_ip instance_name=node1 http://localhost:None "GET /v1.46/containers/roottestasyncloaddatabases-gw7-node1-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node1 http://localhost:None "GET /v1.46/containers/roottestasyncloaddatabases-gw7-node1-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node1, ip: 172.16.2.2... http://localhost:None "GET /v1.46/containers/roottestasyncloaddatabases-gw7-node1-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/77a11e1cce8bd028582c6c4958a4c1383295d41565325da68f4c26ac38ba5e21/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/ed1c3dad61920fe13a676bda7d47ba57d1baa9a4198dbe901f7031f9d256afa3/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/77a11e1cce8bd028582c6c4958a4c1383295d41565325da68f4c26ac38ba5e21/json HTTP/1.1" 200 None Stderr: Network roottestdictionariesalllayoutsseparatesourcesclickhouselocal-gw5_default Creating Stderr: Network roottestdictionariesalllayoutsseparatesourcesclickhouselocal-gw5_default Created Stderr: Container roottestdictionariesalllayoutsseparatesourcesclickhouselocal-gw5-local_node-1 Creating Stderr: Container roottestdictionariesalllayoutsseparatesourcesclickhouselocal-gw5-local_node-1 Created Stderr: Container roottestdictionariesalllayoutsseparatesourcesclickhouselocal-gw5-local_node-1 Starting Stderr: Container roottestdictionariesalllayoutsseparatesourcesclickhouselocal-gw5-local_node-1 Started ClickHouse instance created get_instance_ip instance_name=local_node http://localhost:None "GET /v1.46/containers/roottestdictionariesalllayoutsseparatesourcesclickhouselocal-gw5-local_node-1/json HTTP/1.1" 200 None get_instance_ip instance_name=local_node http://localhost:None "GET /v1.46/containers/roottestdictionariesalllayoutsseparatesourcesclickhouselocal-gw5-local_node-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in local_node, ip: 172.16.3.2... http://localhost:None "GET /v1.46/containers/roottestdictionariesalllayoutsseparatesourcesclickhouselocal-gw5-local_node-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/e66769a1867e3300163adb6fe2d1fe31cf592c8e6703caebefd00470da098014/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/ed1c3dad61920fe13a676bda7d47ba57d1baa9a4198dbe901f7031f9d256afa3/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/77a11e1cce8bd028582c6c4958a4c1383295d41565325da68f4c26ac38ba5e21/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/e66769a1867e3300163adb6fe2d1fe31cf592c8e6703caebefd00470da098014/json HTTP/1.1" 200 None Stderr: Network roottestallowfeaturetier-gw8_default Creating Stderr: Network roottestallowfeaturetier-gw8_default Created Stderr: Container roottestallowfeaturetier-gw8-instance-1 Creating Stderr: Container roottestallowfeaturetier-gw8-instance-1 Created Stderr: Container roottestallowfeaturetier-gw8-instance-1 Starting Stderr: Container roottestallowfeaturetier-gw8-instance-1 Started ClickHouse instance created get_instance_ip instance_name=instance http://localhost:None "GET /v1.46/containers/roottestallowfeaturetier-gw8-instance-1/json HTTP/1.1" 200 None get_instance_ip instance_name=instance http://localhost:None "GET /v1.46/containers/roottestallowfeaturetier-gw8-instance-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in instance, ip: 172.16.4.2... http://localhost:None "GET /v1.46/containers/roottestallowfeaturetier-gw8-instance-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/dc1a70c91e32b7ec4c94d4f19fd5ace4cdfed576c3da9787d19851e2f171ab3c/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/ed1c3dad61920fe13a676bda7d47ba57d1baa9a4198dbe901f7031f9d256afa3/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/77a11e1cce8bd028582c6c4958a4c1383295d41565325da68f4c26ac38ba5e21/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/e66769a1867e3300163adb6fe2d1fe31cf592c8e6703caebefd00470da098014/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/dc1a70c91e32b7ec4c94d4f19fd5ace4cdfed576c3da9787d19851e2f171ab3c/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/ed1c3dad61920fe13a676bda7d47ba57d1baa9a4198dbe901f7031f9d256afa3/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/77a11e1cce8bd028582c6c4958a4c1383295d41565325da68f4c26ac38ba5e21/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/e66769a1867e3300163adb6fe2d1fe31cf592c8e6703caebefd00470da098014/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/dc1a70c91e32b7ec4c94d4f19fd5ace4cdfed576c3da9787d19851e2f171ab3c/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/ed1c3dad61920fe13a676bda7d47ba57d1baa9a4198dbe901f7031f9d256afa3/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/77a11e1cce8bd028582c6c4958a4c1383295d41565325da68f4c26ac38ba5e21/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/e66769a1867e3300163adb6fe2d1fe31cf592c8e6703caebefd00470da098014/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/dc1a70c91e32b7ec4c94d4f19fd5ace4cdfed576c3da9787d19851e2f171ab3c/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/ed1c3dad61920fe13a676bda7d47ba57d1baa9a4198dbe901f7031f9d256afa3/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/77a11e1cce8bd028582c6c4958a4c1383295d41565325da68f4c26ac38ba5e21/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/e66769a1867e3300163adb6fe2d1fe31cf592c8e6703caebefd00470da098014/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/dc1a70c91e32b7ec4c94d4f19fd5ace4cdfed576c3da9787d19851e2f171ab3c/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/ed1c3dad61920fe13a676bda7d47ba57d1baa9a4198dbe901f7031f9d256afa3/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/77a11e1cce8bd028582c6c4958a4c1383295d41565325da68f4c26ac38ba5e21/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/e66769a1867e3300163adb6fe2d1fe31cf592c8e6703caebefd00470da098014/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/dc1a70c91e32b7ec4c94d4f19fd5ace4cdfed576c3da9787d19851e2f171ab3c/json HTTP/1.1" 200 None Stderr:time="2025-04-02T02:18:02Z" level=trace msg="Docker Desktop integration not enabled" Stderr: Network roottestbackuprestoreoncluster-gw2_default Creating Stderr: Network roottestbackuprestoreoncluster-gw2_default Created Stderr: Container roottestbackuprestoreoncluster-gw2-zoo2-1 Creating Stderr: Container roottestbackuprestoreoncluster-gw2-zoo3-1 Creating Stderr: Container roottestbackuprestoreoncluster-gw2-zoo1-1 Creating Stderr: Container roottestbackuprestoreoncluster-gw2-zoo1-1 Created Stderr: Container roottestbackuprestoreoncluster-gw2-zoo3-1 Created Stderr: Container roottestbackuprestoreoncluster-gw2-zoo2-1 Created Stderr: Container roottestbackuprestoreoncluster-gw2-zoo1-1 Starting http://localhost:None "GET /v1.46/containers/ed1c3dad61920fe13a676bda7d47ba57d1baa9a4198dbe901f7031f9d256afa3/json HTTP/1.1" 200 None Stderr: Container roottestbackuprestoreoncluster-gw2-zoo2-1 Starting Stderr: Container roottestbackuprestoreoncluster-gw2-zoo3-1 Starting Stderr: Container roottestbackuprestoreoncluster-gw2-zoo1-1 Started Stderr: Container roottestbackuprestoreoncluster-gw2-zoo3-1 Started Stderr: Container roottestbackuprestoreoncluster-gw2-zoo2-1 Started Stderr:time="2025-04-02T02:18:03Z" level=debug msg="otel error" error="" Stderr:time="2025-04-02T02:18:03Z" level=debug msg="otel error" error="" Wait ZooKeeper to start get_instance_ip instance_name=zoo1 Stderr: Network roottestclusterallreplicas-gw4_default Creating Stderr: Network roottestclusterallreplicas-gw4_default Created Stderr: Container roottestclusterallreplicas-gw4-node1-1 Creating Stderr: Container roottestclusterallreplicas-gw4-node2-1 Creating Stderr: Container roottestclusterallreplicas-gw4-node2-1 Created Stderr: Container roottestclusterallreplicas-gw4-node1-1 Created Stderr: Container roottestclusterallreplicas-gw4-node1-1 Starting Stderr: Container roottestclusterallreplicas-gw4-node2-1 Starting Stderr: Container roottestclusterallreplicas-gw4-node2-1 Started Stderr: Container roottestclusterallreplicas-gw4-node1-1 Started ClickHouse instance created get_instance_ip instance_name=node1 http://localhost:None "GET /v1.46/containers/77a11e1cce8bd028582c6c4958a4c1383295d41565325da68f4c26ac38ba5e21/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottestclusterallreplicas-gw4-node1-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottestbackuprestoreoncluster-gw2-zoo1-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node1 get_kazoo_client: zoo1, ip:172.16.5.2, port:2181, use_ssl:False http://localhost:None "GET /v1.46/containers/roottestclusterallreplicas-gw4-node1-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node1, ip: 172.16.6.3... http://localhost:None "GET /v1.46/containers/roottestclusterallreplicas-gw4-node1-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None Connecting to 172.16.5.2(172.16.5.2):2181, use_ssl: False Connection dropped: socket connection error: Connection refused http://localhost:None "GET /v1.46/containers/e66769a1867e3300163adb6fe2d1fe31cf592c8e6703caebefd00470da098014/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/dc1a70c91e32b7ec4c94d4f19fd5ace4cdfed576c3da9787d19851e2f171ab3c/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/ed1c3dad61920fe13a676bda7d47ba57d1baa9a4198dbe901f7031f9d256afa3/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/77a11e1cce8bd028582c6c4958a4c1383295d41565325da68f4c26ac38ba5e21/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None Connecting to 172.16.5.2(172.16.5.2):2181, use_ssl: False Connection dropped: socket connection error: Connection refused http://localhost:None "GET /v1.46/containers/e66769a1867e3300163adb6fe2d1fe31cf592c8e6703caebefd00470da098014/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/dc1a70c91e32b7ec4c94d4f19fd5ace4cdfed576c3da9787d19851e2f171ab3c/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/ed1c3dad61920fe13a676bda7d47ba57d1baa9a4198dbe901f7031f9d256afa3/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/77a11e1cce8bd028582c6c4958a4c1383295d41565325da68f4c26ac38ba5e21/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/e66769a1867e3300163adb6fe2d1fe31cf592c8e6703caebefd00470da098014/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/dc1a70c91e32b7ec4c94d4f19fd5ace4cdfed576c3da9787d19851e2f171ab3c/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/ed1c3dad61920fe13a676bda7d47ba57d1baa9a4198dbe901f7031f9d256afa3/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/77a11e1cce8bd028582c6c4958a4c1383295d41565325da68f4c26ac38ba5e21/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None Connecting to 172.16.5.2(172.16.5.2):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Stderr:time="2025-04-02T02:18:02Z" level=trace msg="Docker Desktop integration not enabled" Stderr: Network roottestconfigsubstitutions-gw6_default Creating Stderr: Network roottestconfigsubstitutions-gw6_default Created Stderr: Container roottestconfigsubstitutions-gw6-zoo3-1 Creating Stderr: Container roottestconfigsubstitutions-gw6-zoo1-1 Creating Stderr: Container roottestconfigsubstitutions-gw6-zoo2-1 Creating Stderr: Container roottestconfigsubstitutions-gw6-zoo1-1 Created Stderr: Container roottestconfigsubstitutions-gw6-zoo3-1 Created Stderr: Container roottestconfigsubstitutions-gw6-zoo2-1 Created Stderr: Container roottestconfigsubstitutions-gw6-zoo2-1 Starting Stderr: Container roottestconfigsubstitutions-gw6-zoo3-1 Starting Stderr: Container roottestconfigsubstitutions-gw6-zoo1-1 Starting Stderr: Container roottestconfigsubstitutions-gw6-zoo2-1 Started Stderr: Container roottestconfigsubstitutions-gw6-zoo3-1 Started Stderr: Container roottestconfigsubstitutions-gw6-zoo1-1 Started Stderr:time="2025-04-02T02:18:04Z" level=debug msg="otel error" error="" Stderr:time="2025-04-02T02:18:04Z" level=debug msg="otel error" error="" Wait ZooKeeper to start get_instance_ip instance_name=zoo1 http://localhost:None "GET /v1.46/containers/roottestconfigsubstitutions-gw6-zoo1-1/json HTTP/1.1" 200 None get_kazoo_client: zoo1, ip:172.16.7.4, port:2181, use_ssl:False Connecting to 172.16.7.4(172.16.7.4):2181, use_ssl: False Connection dropped: socket connection error: Connection refused http://localhost:None "GET /v1.46/containers/e66769a1867e3300163adb6fe2d1fe31cf592c8e6703caebefd00470da098014/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/dc1a70c91e32b7ec4c94d4f19fd5ace4cdfed576c3da9787d19851e2f171ab3c/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/ed1c3dad61920fe13a676bda7d47ba57d1baa9a4198dbe901f7031f9d256afa3/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/77a11e1cce8bd028582c6c4958a4c1383295d41565325da68f4c26ac38ba5e21/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/e66769a1867e3300163adb6fe2d1fe31cf592c8e6703caebefd00470da098014/json HTTP/1.1" 200 None Connecting to 172.16.7.4(172.16.7.4):2181, use_ssl: False http://localhost:None "GET /v1.46/containers/dc1a70c91e32b7ec4c94d4f19fd5ace4cdfed576c3da9787d19851e2f171ab3c/json HTTP/1.1" 200 None Connection dropped: socket connection error: Connection refused Stderr: proxy2 Skipped - Image is already being pulled by proxy1 Stderr: minio1 Pulling Stderr: instance Pulling Stderr: proxy1 Pulling Stderr: resolver Pulling Stderr: resolver Pulled Stderr: minio1 Pulled Stderr: instance Pulled Stderr: proxy1 Pulled Trying to create Minio instance by command docker compose --project-name roottestdatabasebackup-gw1 --env-file /ClickHouse/tests/integration/test_database_backup/_instances-0-gw1/.env --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_minio.yml --verbose up -d Command:[docker compose --project-name roottestdatabasebackup-gw1 --env-file /ClickHouse/tests/integration/test_database_backup/_instances-0-gw1/.env --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_minio.yml --verbose up -d] http://localhost:None "GET /v1.46/containers/ed1c3dad61920fe13a676bda7d47ba57d1baa9a4198dbe901f7031f9d256afa3/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/77a11e1cce8bd028582c6c4958a4c1383295d41565325da68f4c26ac38ba5e21/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/e66769a1867e3300163adb6fe2d1fe31cf592c8e6703caebefd00470da098014/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/dc1a70c91e32b7ec4c94d4f19fd5ace4cdfed576c3da9787d19851e2f171ab3c/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/ed1c3dad61920fe13a676bda7d47ba57d1baa9a4198dbe901f7031f9d256afa3/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/77a11e1cce8bd028582c6c4958a4c1383295d41565325da68f4c26ac38ba5e21/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/e66769a1867e3300163adb6fe2d1fe31cf592c8e6703caebefd00470da098014/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/dc1a70c91e32b7ec4c94d4f19fd5ace4cdfed576c3da9787d19851e2f171ab3c/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/ed1c3dad61920fe13a676bda7d47ba57d1baa9a4198dbe901f7031f9d256afa3/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/77a11e1cce8bd028582c6c4958a4c1383295d41565325da68f4c26ac38ba5e21/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/e66769a1867e3300163adb6fe2d1fe31cf592c8e6703caebefd00470da098014/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/dc1a70c91e32b7ec4c94d4f19fd5ace4cdfed576c3da9787d19851e2f171ab3c/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/ed1c3dad61920fe13a676bda7d47ba57d1baa9a4198dbe901f7031f9d256afa3/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/77a11e1cce8bd028582c6c4958a4c1383295d41565325da68f4c26ac38ba5e21/json HTTP/1.1" 200 None Connecting to 172.16.7.4(172.16.7.4):2181, use_ssl: False Connection dropped: socket connection error: Connection refused http://localhost:None "GET /v1.46/containers/e66769a1867e3300163adb6fe2d1fe31cf592c8e6703caebefd00470da098014/json HTTP/1.1" 200 None Connecting to 172.16.5.2(172.16.5.2):2181, use_ssl: False Connection dropped: socket connection error: Connection refused http://localhost:None "GET /v1.46/containers/dc1a70c91e32b7ec4c94d4f19fd5ace4cdfed576c3da9787d19851e2f171ab3c/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/ed1c3dad61920fe13a676bda7d47ba57d1baa9a4198dbe901f7031f9d256afa3/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/77a11e1cce8bd028582c6c4958a4c1383295d41565325da68f4c26ac38ba5e21/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/dc1a70c91e32b7ec4c94d4f19fd5ace4cdfed576c3da9787d19851e2f171ab3c/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/e66769a1867e3300163adb6fe2d1fe31cf592c8e6703caebefd00470da098014/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/ed1c3dad61920fe13a676bda7d47ba57d1baa9a4198dbe901f7031f9d256afa3/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/77a11e1cce8bd028582c6c4958a4c1383295d41565325da68f4c26ac38ba5e21/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/e66769a1867e3300163adb6fe2d1fe31cf592c8e6703caebefd00470da098014/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/dc1a70c91e32b7ec4c94d4f19fd5ace4cdfed576c3da9787d19851e2f171ab3c/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/ed1c3dad61920fe13a676bda7d47ba57d1baa9a4198dbe901f7031f9d256afa3/json HTTP/1.1" 200 None Stderr: redis1 Pulling Stderr: node Pulling Stderr: node Pulled Stderr: 6e909acdb790 Pulling fs layer http://localhost:None "GET /v1.46/containers/77a11e1cce8bd028582c6c4958a4c1383295d41565325da68f4c26ac38ba5e21/json HTTP/1.1" 200 None Stderr: d9b98f603d6b Pulling fs layer Stderr: 513666b05c81 Pulling fs layer Stderr: 764e7fbfdb05 Pulling fs layer Stderr: 2884ccae34c1 Pulling fs layer Stderr: df4660a56a60 Pulling fs layer Stderr: 4f4fb700ef54 Pulling fs layer Stderr: b138a3179b82 Pulling fs layer Stderr: 2884ccae34c1 Waiting http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None Stderr: df4660a56a60 Waiting Stderr: 4f4fb700ef54 Waiting Stderr: 764e7fbfdb05 Waiting Stderr: 6e909acdb790 Downloading [> ] 290.2kB/28.2MB Stderr: 513666b05c81 Downloading [==================================================>] 873B/873B Stderr: 513666b05c81 Verifying Checksum Stderr: d9b98f603d6b Downloading [==================================================>] 1.102kB/1.102kB Stderr: d9b98f603d6b Download complete Stderr: 513666b05c81 Download complete Stderr: 764e7fbfdb05 Downloading [> ] 21.02kB/1.438MB Stderr: 2884ccae34c1 Downloading [> ] 162.9kB/15.33MB Stderr: 764e7fbfdb05 Downloading [==================================================>] 1.438MB/1.438MB Stderr: 764e7fbfdb05 Verifying Checksum Stderr: 764e7fbfdb05 Download complete Stderr: df4660a56a60 Downloading [==================================================>] 97B/97B Stderr: df4660a56a60 Verifying Checksum Stderr: df4660a56a60 Download complete Stderr: 4f4fb700ef54 Download complete Stderr: b138a3179b82 Downloading [==================================================>] 572B/572B Stderr: b138a3179b82 Verifying Checksum Stderr: b138a3179b82 Download complete Stderr: 2884ccae34c1 Download complete Stderr: 6e909acdb790 Verifying Checksum Stderr: 6e909acdb790 Download complete Stderr: 6e909acdb790 Extracting [> ] 294.9kB/28.2MB Stderr: 6e909acdb790 Extracting [=========> ] 5.603MB/28.2MB Stderr: 6e909acdb790 Extracting [================> ] 9.437MB/28.2MB Stderr: 6e909acdb790 Extracting [=====================> ] 12.09MB/28.2MB Stderr: 6e909acdb790 Extracting [===============================> ] 17.69MB/28.2MB Stderr: 6e909acdb790 Extracting [======================================> ] 21.82MB/28.2MB Stderr: 6e909acdb790 Extracting [==========================================> ] 23.89MB/28.2MB Stderr: 6e909acdb790 Extracting [===============================================> ] 26.84MB/28.2MB Stderr: 6e909acdb790 Extracting [================================================> ] 27.13MB/28.2MB Stderr: 6e909acdb790 Extracting [================================================> ] 27.43MB/28.2MB Stderr: 6e909acdb790 Extracting [==================================================>] 28.2MB/28.2MB Stderr: 6e909acdb790 Pull complete Stderr: d9b98f603d6b Extracting [==================================================>] 1.102kB/1.102kB Stderr: d9b98f603d6b Extracting [==================================================>] 1.102kB/1.102kB Stderr: d9b98f603d6b Pull complete Stderr: 513666b05c81 Extracting [==================================================>] 873B/873B Stderr: 513666b05c81 Extracting [==================================================>] 873B/873B Stderr: 513666b05c81 Pull complete Stderr: 764e7fbfdb05 Extracting [=> ] 32.77kB/1.438MB Stderr: 764e7fbfdb05 Extracting [==================================================>] 1.438MB/1.438MB Stderr: 764e7fbfdb05 Extracting [==================================================>] 1.438MB/1.438MB Stderr: 764e7fbfdb05 Pull complete Stderr: 2884ccae34c1 Extracting [> ] 163.8kB/15.33MB Stderr: 2884ccae34c1 Extracting [==================> ] 5.571MB/15.33MB Stderr: 2884ccae34c1 Extracting [==============================> ] 9.503MB/15.33MB Stderr: 2884ccae34c1 Extracting [============================================> ] 13.6MB/15.33MB Stderr: 2884ccae34c1 Extracting [==================================================>] 15.33MB/15.33MB Stderr: 2884ccae34c1 Pull complete Stderr: df4660a56a60 Extracting [==================================================>] 97B/97B Stderr: df4660a56a60 Extracting [==================================================>] 97B/97B Stderr: df4660a56a60 Pull complete Stderr: 4f4fb700ef54 Extracting [==================================================>] 32B/32B Stderr: 4f4fb700ef54 Extracting [==================================================>] 32B/32B Stderr: 4f4fb700ef54 Pull complete Stderr: b138a3179b82 Extracting [==================================================>] 572B/572B Stderr: b138a3179b82 Extracting [==================================================>] 572B/572B Stderr: b138a3179b82 Pull complete Stderr: redis1 Pulled Setup Redis Command:[docker compose --project-name roottestdictionariesredis-gw0 --env-file /ClickHouse/tests/integration/test_dictionaries_redis/_instances-0-gw0/.env --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_redis.yml --verbose up -d] http://localhost:None "GET /v1.46/containers/e66769a1867e3300163adb6fe2d1fe31cf592c8e6703caebefd00470da098014/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/dc1a70c91e32b7ec4c94d4f19fd5ace4cdfed576c3da9787d19851e2f171ab3c/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/ed1c3dad61920fe13a676bda7d47ba57d1baa9a4198dbe901f7031f9d256afa3/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/77a11e1cce8bd028582c6c4958a4c1383295d41565325da68f4c26ac38ba5e21/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/e66769a1867e3300163adb6fe2d1fe31cf592c8e6703caebefd00470da098014/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/dc1a70c91e32b7ec4c94d4f19fd5ace4cdfed576c3da9787d19851e2f171ab3c/json HTTP/1.1" 200 None Connecting to 172.16.7.4(172.16.7.4):2181, use_ssl: False Connection dropped: socket connection error: Connection refused http://localhost:None "GET /v1.46/containers/ed1c3dad61920fe13a676bda7d47ba57d1baa9a4198dbe901f7031f9d256afa3/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/77a11e1cce8bd028582c6c4958a4c1383295d41565325da68f4c26ac38ba5e21/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/dc1a70c91e32b7ec4c94d4f19fd5ace4cdfed576c3da9787d19851e2f171ab3c/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/e66769a1867e3300163adb6fe2d1fe31cf592c8e6703caebefd00470da098014/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/ed1c3dad61920fe13a676bda7d47ba57d1baa9a4198dbe901f7031f9d256afa3/json HTTP/1.1" 200 None ClickHouse node started Executing query CREATE DATABASE test ENGINE = Ordinary on node http://localhost:None "GET /v1.46/containers/77a11e1cce8bd028582c6c4958a4c1383295d41565325da68f4c26ac38ba5e21/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/dc1a70c91e32b7ec4c94d4f19fd5ace4cdfed576c3da9787d19851e2f171ab3c/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/e66769a1867e3300163adb6fe2d1fe31cf592c8e6703caebefd00470da098014/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/77a11e1cce8bd028582c6c4958a4c1383295d41565325da68f4c26ac38ba5e21/json HTTP/1.1" 200 None ClickHouse node1 started get_instance_ip instance_name=node2 http://localhost:None "GET /v1.46/containers/roottestasyncloaddatabases-gw7-node2-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node2 http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottestasyncloaddatabases-gw7-node2-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node2, ip: 172.16.2.3... http://localhost:None "GET /v1.46/containers/roottestasyncloaddatabases-gw7-node2-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/ce24de9b97f22c82bd0b2b3a77bacac6829d2f5fe090326c8a83c77a8595374e/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/dc1a70c91e32b7ec4c94d4f19fd5ace4cdfed576c3da9787d19851e2f171ab3c/json HTTP/1.1" 200 None Stderr:time="2025-04-02T02:18:04Z" level=trace msg="Docker Desktop integration not enabled" Stderr: Network roottestdatabasebackup-gw1_default Creating Stderr: Network roottestdatabasebackup-gw1_default Created Stderr: Volume "roottestdatabasebackup-gw1_data1-1" Creating Stderr: Volume "roottestdatabasebackup-gw1_data1-1" Created Stderr: Container roottestdatabasebackup-gw1-proxy2-1 Creating http://localhost:None "GET /v1.46/containers/e66769a1867e3300163adb6fe2d1fe31cf592c8e6703caebefd00470da098014/json HTTP/1.1" 200 None Stderr: Container roottestdatabasebackup-gw1-proxy1-1 Creating Stderr: Container roottestdatabasebackup-gw1-proxy1-1 Created Stderr: Container roottestdatabasebackup-gw1-proxy2-1 Created Stderr: Container roottestdatabasebackup-gw1-minio1-1 Creating Stderr: Container roottestdatabasebackup-gw1-resolver-1 Creating Stderr: Container roottestdatabasebackup-gw1-resolver-1 Created Stderr: Container roottestdatabasebackup-gw1-minio1-1 Created Stderr: Container roottestdatabasebackup-gw1-proxy2-1 Starting Stderr: Container roottestdatabasebackup-gw1-proxy1-1 Starting Stderr: Container roottestdatabasebackup-gw1-proxy2-1 Started Stderr: Container roottestdatabasebackup-gw1-proxy1-1 Started Stderr: Container roottestdatabasebackup-gw1-minio1-1 Starting Stderr: Container roottestdatabasebackup-gw1-resolver-1 Starting Stderr: Container roottestdatabasebackup-gw1-minio1-1 Started Stderr: Container roottestdatabasebackup-gw1-resolver-1 Started Stderr:time="2025-04-02T02:18:05Z" level=debug msg="otel error" error="" Stderr:time="2025-04-02T02:18:05Z" level=debug msg="otel error" error="" Trying to connect to Minio... get_instance_ip instance_name=minio1 http://localhost:None "GET /v1.46/containers/roottestdatabasebackup-gw1-minio1-1/json HTTP/1.1" 200 None get_instance_ip instance_name=proxy1 http://localhost:None "GET /v1.46/containers/roottestdatabasebackup-gw1-proxy1-1/json HTTP/1.1" 200 None Starting new HTTP connection (1): 172.16.8.4:9001 Incremented Retry for (url='/'): Retry(total=2, connect=None, read=None, redirect=None, status=None) Retrying (Retry(total=2, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError(': Failed to establish a new connection: [Errno 111] Connection refused')': / Starting new HTTP connection (2): 172.16.8.4:9001 Incremented Retry for (url='/'): Retry(total=1, connect=None, read=None, redirect=None, status=None) Retrying (Retry(total=1, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError(': Failed to establish a new connection: [Errno 111] Connection refused')': / Starting new HTTP connection (3): 172.16.8.4:9001 Incremented Retry for (url='/'): Retry(total=0, connect=None, read=None, redirect=None, status=None) Retrying (Retry(total=0, connect=None, read=None, redirect=None, status=None)) after connection broken by 'NewConnectionError(': Failed to establish a new connection: [Errno 111] Connection refused')': / Starting new HTTP connection (4): 172.16.8.4:9001 Can't connect to Minio: HTTPConnectionPool(host='172.16.8.4', port=9001): Max retries exceeded with url: / (Caused by NewConnectionError(': Failed to establish a new connection: [Errno 111] Connection refused')) http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/ce24de9b97f22c82bd0b2b3a77bacac6829d2f5fe090326c8a83c77a8595374e/json HTTP/1.1" 200 None Executing query DROP TABLE IF EXISTS test.tbl on node http://localhost:None "GET /v1.46/containers/dc1a70c91e32b7ec4c94d4f19fd5ace4cdfed576c3da9787d19851e2f171ab3c/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/e66769a1867e3300163adb6fe2d1fe31cf592c8e6703caebefd00470da098014/json HTTP/1.1" 200 None ClickHouse local_node started Executing query CREATE DATABASE IF NOT EXISTS test on local_node http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/ce24de9b97f22c82bd0b2b3a77bacac6829d2f5fe090326c8a83c77a8595374e/json HTTP/1.1" 200 None Stderr:time="2025-04-02T02:18:05Z" level=trace msg="Docker Desktop integration not enabled" Stderr: Network roottestdictionariesredis-gw0_default Creating Stderr: Network roottestdictionariesredis-gw0_default Created Stderr: Container roottestdictionariesredis-gw0-redis1-1 Creating Stderr: Container roottestdictionariesredis-gw0-redis1-1 Created Stderr: Container roottestdictionariesredis-gw0-redis1-1 Starting Stderr: Container roottestdictionariesredis-gw0-redis1-1 Started Stderr:time="2025-04-02T02:18:05Z" level=debug msg="otel error" error="" Stderr:time="2025-04-02T02:18:05Z" level=debug msg="otel error" error="" http://localhost:None "GET /v1.46/containers/dc1a70c91e32b7ec4c94d4f19fd5ace4cdfed576c3da9787d19851e2f171ab3c/json HTTP/1.1" 200 None ClickHouse instance started Executing query SELECT value FROM system.server_settings where name = 'allow_feature_tier' on instance http://localhost:None "GET /v1.46/containers/ce24de9b97f22c82bd0b2b3a77bacac6829d2f5fe090326c8a83c77a8595374e/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None Executing query CREATE TABLE test.tbl (p Date, k Int8) ENGINE = MergeTree PARTITION BY toYYYYMM(p) ORDER BY p on node Executing query CREATE TABLE test.table_LocalClickHouse_flat_ ( KeyField UInt64,UInt8_ UInt8,UInt16_ UInt16,UInt32_ UInt32,UInt64_ UInt64,Int8_ Int8,Int16_ Int16,Int32_ Int32,Int64_ Int64,UUID_ UUID,Date_ Date,DateTime_ DateTime,String_ String,Float32_ Float32,Float64_ Float64,ParentKeyField UInt64) ENGINE MergeTree ORDER BY tuple(); on local_node http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/ce24de9b97f22c82bd0b2b3a77bacac6829d2f5fe090326c8a83c77a8595374e/json HTTP/1.1" 200 None ClickHouse node2 started Executing query CREATE DATABASE IF NOT EXISTS dict ENGINE=Dictionary; CREATE DATABASE IF NOT EXISTS test; on node1 Connecting to 172.16.7.4(172.16.7.4):2181, use_ssl: False Connection dropped: socket connection error: Connection refused http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None Executing query INSERT INTO test.tbl (p, k) VALUES(toDate(1), 1) on node Executing query SELECT 1 SETTINGS allow_experimental_time_series_table=1 on instance http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None Executing query INSERT INTO test.table_LocalClickHouse_flat_ (KeyField,UInt8_,UInt16_,UInt32_,UInt64_,Int8_,Int16_,Int32_,Int64_,UUID_,Date_,DateTime_,String_,Float32_,Float64_,ParentKeyField) values (1,22,333,4444,55555,-6,-77,-888,-999,'550e8400-e29b-41d4-a716-446655440003','1973-06-28','1985-02-28 23:43:25','hello',22.543,3332154213.4,0),(2,3,4,5,6,-7,-8,-9,-10,'550e8400-e29b-41d4-a716-446655440002','1978-06-28','1986-02-28 23:42:25','hello',21.543,3222154213.4,1) on local_node Executing query CREATE DATABASE IF NOT EXISTS dict ENGINE=Dictionary; CREATE DATABASE IF NOT EXISTS test; on node2 http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None Executing query INSERT INTO test.tbl (p, k) VALUES(toDate(2), 2) on node run container_id:roottestallowfeaturetier-gw8-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "sed -i 's/0/1/g' /etc/clickhouse-server/config.d/allow_feature_tier.xml"] Command:[docker exec roottestallowfeaturetier-gw8-instance-1 bash -c sed -i 's/0/1/g' /etc/clickhouse-server/config.d/allow_feature_tier.xml] http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None Connecting to 172.16.5.2(172.16.5.2):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Executing query system flush logs on node2 Executing query SYSTEM RELOAD CONFIG on instance Executing query CREATE DATABASE IF NOT EXISTS test on local_node http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None Executing query INSERT INTO test.tbl (p, k) VALUES(toDate(3), 3) on node Starting new HTTP connection (5): 172.16.8.4:9001 Executing query CREATE TABLE test.table_LocalClickHouse_hashed_ ( KeyField UInt64,UInt8_ UInt8,UInt16_ UInt16,UInt32_ UInt32,UInt64_ UInt64,Int8_ Int8,Int16_ Int16,Int32_ Int32,Int64_ Int64,UUID_ UUID,Date_ Date,DateTime_ DateTime,String_ String,Float32_ Float32,Float64_ Float64,ParentKeyField UInt64) ENGINE MergeTree ORDER BY tuple(); on local_node http://172.16.8.4:9001 "GET / HTTP/1.1" 200 0 Connected to Minio. http://172.16.8.4:9001 "GET /root?location= HTTP/1.1" 404 0 http://172.16.8.4:9001 "PUT /root HTTP/1.1" 200 0 S3 bucket 'root' created http://172.16.8.4:9001 "GET /root2?location= HTTP/1.1" 404 0 http://172.16.8.4:9001 "PUT /root2 HTTP/1.1" 200 0 S3 bucket 'root2' created ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_database_backup/_instances-0-gw1/.env --project-name roottestdatabasebackup-gw1 --file /ClickHouse/tests/integration/test_database_backup/_instances-0-gw1/instance/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_minio.yml up -d --no-recreate') Command:[docker compose --env-file /ClickHouse/tests/integration/test_database_backup/_instances-0-gw1/.env --project-name roottestdatabasebackup-gw1 --file /ClickHouse/tests/integration/test_database_backup/_instances-0-gw1/instance/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_minio.yml up -d --no-recreate] http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None Executing query SELECT value FROM system.server_settings where name = 'allow_feature_tier' on instance http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None Executing query INSERT INTO test.table_LocalClickHouse_hashed_ (KeyField,UInt8_,UInt16_,UInt32_,UInt64_,Int8_,Int16_,Int32_,Int64_,UUID_,Date_,DateTime_,String_,Float32_,Float64_,ParentKeyField) values (1,22,333,4444,55555,-6,-77,-888,-999,'550e8400-e29b-41d4-a716-446655440003','1973-06-28','1985-02-28 23:43:25','hello',22.543,3332154213.4,0),(2,3,4,5,6,-7,-8,-9,-10,'550e8400-e29b-41d4-a716-446655440002','1978-06-28','1986-02-28 23:42:25','hello',21.543,3222154213.4,1) on local_node Executing query INSERT INTO test.tbl (p, k) VALUES(toDate(31), 31) on node http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None Executing query SELECT 1 SETTINGS allow_experimental_time_series_table=1 on instance http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None Stderr: Container roottestdatabasebackup-gw1-proxy2-1 Running Stderr: Container roottestdatabasebackup-gw1-proxy1-1 Running Stderr: Container roottestdatabasebackup-gw1-minio1-1 Running Stderr: Container roottestdatabasebackup-gw1-instance-1 Creating Stderr: Container roottestdatabasebackup-gw1-resolver-1 Running Stderr: Container roottestdatabasebackup-gw1-instance-1 Created Stderr: Container roottestdatabasebackup-gw1-instance-1 Starting Stderr: Container roottestdatabasebackup-gw1-instance-1 Started ClickHouse instance created get_instance_ip instance_name=instance http://localhost:None "GET /v1.46/containers/roottestdatabasebackup-gw1-instance-1/json HTTP/1.1" 200 None get_instance_ip instance_name=instance http://localhost:None "GET /v1.46/containers/roottestdatabasebackup-gw1-instance-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in instance, ip: 172.16.8.6... http://localhost:None "GET /v1.46/containers/roottestdatabasebackup-gw1-instance-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/49d34b1d947af59f688878be35f5f1fbcaccc16760a3f92d90a5a4557b52c688/json HTTP/1.1" 200 None Executing query CREATE DATABASE IF NOT EXISTS test on local_node http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/49d34b1d947af59f688878be35f5f1fbcaccc16760a3f92d90a5a4557b52c688/json HTTP/1.1" 200 None Executing query INSERT INTO test.tbl (p, k) VALUES(toDate(32), 32) on node http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/49d34b1d947af59f688878be35f5f1fbcaccc16760a3f92d90a5a4557b52c688/json HTTP/1.1" 200 None Executing query CREATE TABLE test.table_LocalClickHouse_cache_ ( KeyField UInt64,UInt8_ UInt8,UInt16_ UInt16,UInt32_ UInt32,UInt64_ UInt64,Int8_ Int8,Int16_ Int16,Int32_ Int32,Int64_ Int64,UUID_ UUID,Date_ Date,DateTime_ DateTime,String_ String,Float32_ Float32,Float64_ Float64,ParentKeyField UInt64) ENGINE MergeTree ORDER BY tuple(); on local_node Executing query SELECT 1 SETTINGS enable_parallel_replicas=1 on instance Connecting to 172.16.7.4(172.16.7.4):2181, use_ssl: False Connection dropped: socket connection error: Connection refused http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/49d34b1d947af59f688878be35f5f1fbcaccc16760a3f92d90a5a4557b52c688/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None Executing query select count() from system.tables where database = 'system' and name in ['query_log', 'text_log'] on node2 Executing query INSERT INTO test.tbl (p, k) VALUES(toDate(33), 33) on node http://localhost:None "GET /v1.46/containers/49d34b1d947af59f688878be35f5f1fbcaccc16760a3f92d90a5a4557b52c688/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None Executing query INSERT INTO test.table_LocalClickHouse_cache_ (KeyField,UInt8_,UInt16_,UInt32_,UInt64_,Int8_,Int16_,Int32_,Int64_,UUID_,Date_,DateTime_,String_,Float32_,Float64_,ParentKeyField) values (1,22,333,4444,55555,-6,-77,-888,-999,'550e8400-e29b-41d4-a716-446655440003','1973-06-28','1985-02-28 23:43:25','hello',22.543,3332154213.4,0),(2,3,4,5,6,-7,-8,-9,-10,'550e8400-e29b-41d4-a716-446655440002','1978-06-28','1986-02-28 23:42:25','hello',21.543,3222154213.4,1) on local_node http://localhost:None "GET /v1.46/containers/49d34b1d947af59f688878be35f5f1fbcaccc16760a3f92d90a5a4557b52c688/json HTTP/1.1" 200 None run container_id:roottestallowfeaturetier-gw8-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "sed -i 's/1/2/g' /etc/clickhouse-server/config.d/allow_feature_tier.xml"] Command:[docker exec roottestallowfeaturetier-gw8-instance-1 bash -c sed -i 's/1/2/g' /etc/clickhouse-server/config.d/allow_feature_tier.xml] http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None Executing query SYSTEM RELOAD CONFIG on instance http://localhost:None "GET /v1.46/containers/49d34b1d947af59f688878be35f5f1fbcaccc16760a3f92d90a5a4557b52c688/json HTTP/1.1" 200 None Executing query SELECT * FROM test.tbl ORDER BY p on node http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None Executing query rename table system.text_log to system.text_log_1_test on node2 http://localhost:None "GET /v1.46/containers/49d34b1d947af59f688878be35f5f1fbcaccc16760a3f92d90a5a4557b52c688/json HTTP/1.1" 200 None Executing query CREATE DATABASE IF NOT EXISTS test on local_node http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/49d34b1d947af59f688878be35f5f1fbcaccc16760a3f92d90a5a4557b52c688/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/49d34b1d947af59f688878be35f5f1fbcaccc16760a3f92d90a5a4557b52c688/json HTTP/1.1" 200 None Executing query rename table system.query_log to system.query_log_1_test on node2 Executing query CREATE TABLE test.table_LocalClickHouse_direct_ ( KeyField UInt64,UInt8_ UInt8,UInt16_ UInt16,UInt32_ UInt32,UInt64_ UInt64,Int8_ Int8,Int16_ Int16,Int32_ Int32,Int64_ Int64,UUID_ UUID,Date_ Date,DateTime_ DateTime,String_ String,Float32_ Float32,Float64_ Float64,ParentKeyField UInt64) ENGINE MergeTree ORDER BY tuple(); on local_node Executing query ALTER TABLE test.tbl FREEZE on node http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None Executing query SELECT value FROM system.server_settings where name = 'allow_feature_tier' on instance http://localhost:None "GET /v1.46/containers/49d34b1d947af59f688878be35f5f1fbcaccc16760a3f92d90a5a4557b52c688/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/49d34b1d947af59f688878be35f5f1fbcaccc16760a3f92d90a5a4557b52c688/json HTTP/1.1" 200 None Executing query system flush logs on node2 Executing query INSERT INTO test.table_LocalClickHouse_direct_ (KeyField,UInt8_,UInt16_,UInt32_,UInt64_,Int8_,Int16_,Int32_,Int64_,UUID_,Date_,DateTime_,String_,Float32_,Float64_,ParentKeyField) values (1,22,333,4444,55555,-6,-77,-888,-999,'550e8400-e29b-41d4-a716-446655440003','1973-06-28','1985-02-28 23:43:25','hello',22.543,3332154213.4,0),(2,3,4,5,6,-7,-8,-9,-10,'550e8400-e29b-41d4-a716-446655440002','1978-06-28','1986-02-28 23:42:25','hello',21.543,3222154213.4,1) on local_node Executing query CREATE TABLE test.tbl2 AS test.tbl on node http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/49d34b1d947af59f688878be35f5f1fbcaccc16760a3f92d90a5a4557b52c688/json HTTP/1.1" 200 None Executing query SELECT 1 SETTINGS allow_experimental_time_series_table=1 on instance http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/49d34b1d947af59f688878be35f5f1fbcaccc16760a3f92d90a5a4557b52c688/json HTTP/1.1" 200 None Executing query INSERT INTO test.tbl2(p, k) VALUES(toDate(3), 3) on node http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/49d34b1d947af59f688878be35f5f1fbcaccc16760a3f92d90a5a4557b52c688/json HTTP/1.1" 200 None Executing query CREATE DATABASE IF NOT EXISTS test on local_node http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None Executing query SELECT 1 SETTINGS enable_parallel_replicas=1 on instance http://localhost:None "GET /v1.46/containers/49d34b1d947af59f688878be35f5f1fbcaccc16760a3f92d90a5a4557b52c688/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/49d34b1d947af59f688878be35f5f1fbcaccc16760a3f92d90a5a4557b52c688/json HTTP/1.1" 200 None Executing query CREATE TABLE test.table_LocalClickHouse_complex_key_hashed_ ( KeyField1 UInt64,KeyField2 String,UInt8_ UInt8,UInt16_ UInt16,UInt32_ UInt32,UInt64_ UInt64,Int8_ Int8,Int16_ Int16,Int32_ Int32,Int64_ Int64,UUID_ UUID,Date_ Date,DateTime_ DateTime,String_ String,Float32_ Float32,Float64_ Float64) ENGINE MergeTree ORDER BY tuple(); on local_node Executing query INSERT INTO test.tbl2(p, k) VALUES(toDate(4), 4) on node http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/49d34b1d947af59f688878be35f5f1fbcaccc16760a3f92d90a5a4557b52c688/json HTTP/1.1" 200 None run container_id:roottestallowfeaturetier-gw8-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "sed -i 's/2/0/g' /etc/clickhouse-server/config.d/allow_feature_tier.xml"] Command:[docker exec roottestallowfeaturetier-gw8-instance-1 bash -c sed -i 's/2/0/g' /etc/clickhouse-server/config.d/allow_feature_tier.xml] http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None Executing query SYSTEM RELOAD CONFIG on instance http://localhost:None "GET /v1.46/containers/49d34b1d947af59f688878be35f5f1fbcaccc16760a3f92d90a5a4557b52c688/json HTTP/1.1" 200 None Executing query INSERT INTO test.table_LocalClickHouse_complex_key_hashed_ (KeyField1,KeyField2,UInt8_,UInt16_,UInt32_,UInt64_,Int8_,Int16_,Int32_,Int64_,UUID_,Date_,DateTime_,String_,Float32_,Float64_) values (1,'world',22,333,4444,55555,-6,-77,-888,-999,'550e8400-e29b-41d4-a716-446655440003','1973-06-28','1985-02-28 23:43:25','hello',22.543,3332154213.4),(2,'qwerty2',52,2345,6544,9191991,-2,-717,-81818,-92929,'550e8400-e29b-41d4-a716-446655440007','1975-09-28','2000-02-28 23:33:24','my',255.543,3332221.44) on local_node http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None Executing query INSERT INTO test.tbl2(p, k) VALUES(toDate(33), 33) on node http://localhost:None "GET /v1.46/containers/49d34b1d947af59f688878be35f5f1fbcaccc16760a3f92d90a5a4557b52c688/json HTTP/1.1" 200 None Executing query select count() from system.tables where database = 'system' and name in ['query_log', 'text_log'] on node2 http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/49d34b1d947af59f688878be35f5f1fbcaccc16760a3f92d90a5a4557b52c688/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None Connecting to 172.16.7.4(172.16.7.4):2181, use_ssl: False Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Zookeeper connection established, state: CONNECTED Sending request(xid=1): GetChildren(path='/', watcher=None) Received response(xid=1): ['keeper'] Sending request(xid=2): Close() Executing query CREATE DATABASE IF NOT EXISTS test on local_node Connection dropped: socket connection broken Transition to CONNECTING Zookeeper connection lost http://localhost:None "GET /v1.46/containers/49d34b1d947af59f688878be35f5f1fbcaccc16760a3f92d90a5a4557b52c688/json HTTP/1.1" 200 None Executing query SELECT value FROM system.server_settings where name = 'allow_feature_tier' on instance Connecting to 172.16.5.2(172.16.5.2):2181, use_ssl: False Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Zookeeper connection established, state: CONNECTED Sending request(xid=1): GetChildren(path='/', watcher=None) Received response(xid=1): ['keeper'] Sending request(xid=2): Close() Connection dropped: socket connection broken Transition to CONNECTING Zookeeper connection lost Executing query INSERT INTO test.tbl2(p, k) VALUES(toDate(34), 34) on node http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None Failed connecting to Zookeeper within the connection retry policy. Zookeeper session closed, state: CLOSED get_instance_ip instance_name=zoo2 http://localhost:None "GET /v1.46/containers/roottestbackuprestoreoncluster-gw2-zoo2-1/json HTTP/1.1" 200 None get_kazoo_client: zoo2, ip:172.16.5.3, port:2181, use_ssl:False Connecting to 172.16.5.3(172.16.5.3):2181, use_ssl: False Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Zookeeper connection established, state: CONNECTED Sending request(xid=1): GetChildren(path='/', watcher=None) Received response(xid=1): ['keeper'] Sending request(xid=2): Close() Failed connecting to Zookeeper within the connection retry policy. Zookeeper session closed, state: CLOSED get_instance_ip instance_name=zoo2 Connection dropped: socket connection broken Transition to CONNECTING Zookeeper connection lost http://localhost:None "GET /v1.46/containers/roottestconfigsubstitutions-gw6-zoo2-1/json HTTP/1.1" 200 None get_kazoo_client: zoo2, ip:172.16.7.2, port:2181, use_ssl:False Connecting to 172.16.7.2(172.16.7.2):2181, use_ssl: False Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) http://localhost:None "GET /v1.46/containers/49d34b1d947af59f688878be35f5f1fbcaccc16760a3f92d90a5a4557b52c688/json HTTP/1.1" 200 None Zookeeper connection established, state: CONNECTED Sending request(xid=1): GetChildren(path='/', watcher=None) Received response(xid=1): ['keeper'] Sending request(xid=2): Close() Connection dropped: socket connection broken Transition to CONNECTING Zookeeper connection lost Executing query rename table system.text_log to system.text_log_2_test on node2 http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None Failed connecting to Zookeeper within the connection retry policy. Zookeeper session closed, state: CLOSED get_instance_ip instance_name=zoo3 http://localhost:None "GET /v1.46/containers/roottestbackuprestoreoncluster-gw2-zoo3-1/json HTTP/1.1" 200 None get_kazoo_client: zoo3, ip:172.16.5.4, port:2181, use_ssl:False Connecting to 172.16.5.4(172.16.5.4):2181, use_ssl: False Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) http://localhost:None "GET /v1.46/containers/49d34b1d947af59f688878be35f5f1fbcaccc16760a3f92d90a5a4557b52c688/json HTTP/1.1" 200 None ClickHouse instance started run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', 'rm -rf /backups/'] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c rm -rf /backups/] Executing query CREATE TABLE test.table_LocalClickHouse_complex_key_cache_ ( KeyField1 UInt64,KeyField2 String,UInt8_ UInt8,UInt16_ UInt16,UInt32_ UInt32,UInt64_ UInt64,Int8_ Int8,Int16_ Int16,Int32_ Int32,Int64_ Int64,UUID_ UUID,Date_ Date,DateTime_ DateTime,String_ String,Float32_ Float32,Float64_ Float64) ENGINE MergeTree ORDER BY tuple(); on local_node Zookeeper connection established, state: CONNECTED Sending request(xid=1): GetChildren(path='/', watcher=None) Received response(xid=1): ['keeper'] Failed connecting to Zookeeper within the connection retry policy. Sending request(xid=2): Close() Zookeeper session closed, state: CLOSED get_instance_ip instance_name=zoo3 http://localhost:None "GET /v1.46/containers/roottestconfigsubstitutions-gw6-zoo3-1/json HTTP/1.1" 200 None get_kazoo_client: zoo3, ip:172.16.7.3, port:2181, use_ssl:False Connection dropped: socket connection broken Transition to CONNECTING Connecting to 172.16.7.3(172.16.7.3):2181, use_ssl: False Zookeeper connection lost Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Zookeeper connection established, state: CONNECTED Sending request(xid=1): GetChildren(path='/', watcher=None) Received response(xid=1): ['keeper'] Sending request(xid=2): Close() Connection dropped: socket connection broken Transition to CONNECTING Zookeeper connection lost run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', 'rm -rf /local_plain/'] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c rm -rf /local_plain/] http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None http://172.16.8.4:9001 "GET /root?delimiter=&encoding-type=url&list-type=2&max-keys=1000&prefix= HTTP/1.1" 200 0 Executing query DROP DATABASE IF EXISTS test_database; DROP DATABASE IF EXISTS test_database_backup; CREATE DATABASE test_database; CREATE TABLE test_database.test_table_1 (id UInt64, value String) ENGINE=MergeTree ORDER BY id; INSERT INTO test_database.test_table_1 VALUES (0, 'test_database.test_table_1'); CREATE TABLE test_database.test_table_2 (id UInt64, value String) ENGINE=MergeTree ORDER BY id; INSERT INTO test_database.test_table_2 VALUES (0, 'test_database.test_table_2'); CREATE TABLE test_database.test_table_3 (id UInt64, value String) ENGINE=MergeTree ORDER BY id; INSERT INTO test_database.test_table_3 VALUES (0, 'test_database.test_table_3'); BACKUP DATABASE test_database TO Disk('backup_disk_local', 'test_database_backup'); CREATE DATABASE test_database_backup ENGINE = Backup('test_database', Disk('backup_disk_local', 'test_database_backup')); on instance Executing query SELECT * FROM test.tbl2 ORDER BY p on node [gw8] PASSED test_allow_feature_tier/test.py::test_allow_feature_tier_in_general_settings Failed connecting to Zookeeper within the connection retry policy. test_allow_feature_tier/test.py::test_allow_feature_tier_in_mergetree_settings Zookeeper session closed, state: CLOSED All instances of ZooKeeper started: ('zoo1', 'zoo2', 'zoo3') ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw2/.env --project-name roottestbackuprestoreoncluster-gw2 --file /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw2/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw2/node2/docker-compose.yml --file /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw2/node3/docker-compose.yml up -d --no-recreate') Command:[docker compose --env-file /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw2/.env --project-name roottestbackuprestoreoncluster-gw2 --file /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw2/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw2/node2/docker-compose.yml --file /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw2/node3/docker-compose.yml up -d --no-recreate] Executing query SELECT value FROM system.server_settings where name = 'allow_feature_tier' on instance Failed connecting to Zookeeper within the connection retry policy. Zookeeper session closed, state: CLOSED All instances of ZooKeeper started: ('zoo1', 'zoo2', 'zoo3') get_instance_ip instance_name=zoo1 http://localhost:None "GET /v1.46/containers/roottestconfigsubstitutions-gw6-zoo1-1/json HTTP/1.1" 200 None get_kazoo_client: zoo1, ip:172.16.7.4, port:2181, use_ssl:False Connecting to 172.16.7.4(172.16.7.4):2181, use_ssl: False Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Zookeeper connection established, state: CONNECTED run_kazoo_commands_with_retries: zoo1, .create_zk_roots at 0x7fd927c84550> Sending request(xid=1): Create(path='/setting/max_query_size', data=b'77777', acl=[ACL(perms=31, acl_list=['ALL'], id=Id(scheme='world', id='anyone'))], flags=0) Received error(xid=1) NoNodeError() Sending request(xid=2): Exists(path='/setting', watcher=None) Sending request(xid=3): Exists(path='/', watcher=None) Received response(xid=3): ZnodeStat(czxid=0, mzxid=0, ctime=0, mtime=0, version=0, cversion=0, aversion=0, ephemeralOwner=0, dataLength=0, numChildren=1, pzxid=0) Sending request(xid=4): Create(path='/setting', data=b'', acl=[ACL(perms=31, acl_list=['ALL'], id=Id(scheme='world', id='anyone'))], flags=0) Received response(xid=4): '/setting' Sending request(xid=5): Create(path='/setting/max_query_size', data=b'77777', acl=[ACL(perms=31, acl_list=['ALL'], id=Id(scheme='world', id='anyone'))], flags=0) Received response(xid=5): '/setting/max_query_size' Sending request(xid=6): Create(path='/users_from_zk_1', data=b'default', acl=[ACL(perms=31, acl_list=['ALL'], id=Id(scheme='world', id='anyone'))], flags=0) Received response(xid=6): '/users_from_zk_1' Sending request(xid=7): Create(path='/users_from_zk_2', data=b'default', acl=[ACL(perms=31, acl_list=['ALL'], id=Id(scheme='world', id='anyone'))], flags=0) Received response(xid=7): '/users_from_zk_2' Sending request(xid=8): Create(path='/min_bytes_for_wide_part', data=b'33', acl=[ACL(perms=31, acl_list=['ALL'], id=Id(scheme='world', id='anyone'))], flags=0) Received response(xid=8): '/min_bytes_for_wide_part' Sending request(xid=9): Create(path='/merge_max_block_size', data=b'8888', acl=[ACL(perms=31, acl_list=['ALL'], id=Id(scheme='world', id='anyone'))], flags=0) Received response(xid=9): '/merge_max_block_size' ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw6/.env --project-name roottestconfigsubstitutions-gw6 --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw6/node1/docker-compose.yml --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw6/node2/docker-compose.yml --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw6/node3/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw6/node4/docker-compose.yml --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw6/node5/docker-compose.yml --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw6/node6/docker-compose.yml --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw6/node7/docker-compose.yml --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw6/node8/docker-compose.yml up -d --no-recreate') Command:[docker compose --env-file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw6/.env --project-name roottestconfigsubstitutions-gw6 --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw6/node1/docker-compose.yml --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw6/node2/docker-compose.yml --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw6/node3/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw6/node4/docker-compose.yml --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw6/node5/docker-compose.yml --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw6/node6/docker-compose.yml --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw6/node7/docker-compose.yml --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw6/node8/docker-compose.yml up -d --no-recreate] http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None Executing query rename table system.query_log to system.query_log_2_test on node2 Executing query INSERT INTO test.table_LocalClickHouse_complex_key_cache_ (KeyField1,KeyField2,UInt8_,UInt16_,UInt32_,UInt64_,Int8_,Int16_,Int32_,Int64_,UUID_,Date_,DateTime_,String_,Float32_,Float64_) values (1,'world',22,333,4444,55555,-6,-77,-888,-999,'550e8400-e29b-41d4-a716-446655440003','1973-06-28','1985-02-28 23:43:25','hello',22.543,3332154213.4),(2,'qwerty2',52,2345,6544,9191991,-2,-717,-81818,-92929,'550e8400-e29b-41d4-a716-446655440007','1975-09-28','2000-02-28 23:33:24','my',255.543,3332221.44) on local_node http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None copy from /var/lib/clickhouse/shadow/*/data/test/tbl to /var/lib/clickhouse/data/test/tbl2/detached run container_id:roottestbackuprestore-gw9-node-1 detach:False nothrow:False cmd: ['bash', '-c', 'cp -r /var/lib/clickhouse/shadow/*/data/test/tbl -T /var/lib/clickhouse/data/test/tbl2/detached'] Command:[docker exec roottestbackuprestore-gw9-node-1 bash -c cp -r /var/lib/clickhouse/shadow/*/data/test/tbl -T /var/lib/clickhouse/data/test/tbl2/detached] Executing query DROP TABLE IF EXISTS test_experimental on instance Executing query system flush logs on node2 Executing query ALTER TABLE test.tbl2 ATTACH PARTITION 197001 on node http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None Executing query SELECT id, value FROM test_database_backup.test_table_1 on instance Executing query CREATE DATABASE IF NOT EXISTS test on local_node http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None run container_id:roottestallowfeaturetier-gw8-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "sed -i 's/0/1/g' /etc/clickhouse-server/config.d/allow_feature_tier.xml"] Command:[docker exec roottestallowfeaturetier-gw8-instance-1 bash -c sed -i 's/0/1/g' /etc/clickhouse-server/config.d/allow_feature_tier.xml] Executing query SYSTEM RELOAD CONFIG on instance http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None Stderr: Container roottestbackuprestoreoncluster-gw2-zoo1-1 Running Stderr: Container roottestbackuprestoreoncluster-gw2-zoo2-1 Running Stderr: Container roottestbackuprestoreoncluster-gw2-zoo3-1 Running Stderr: Container roottestbackuprestoreoncluster-gw2-node2-1 Creating Stderr: Container roottestbackuprestoreoncluster-gw2-node3-1 Creating Stderr: Container roottestbackuprestoreoncluster-gw2-node1-1 Creating Stderr: Container roottestbackuprestoreoncluster-gw2-node2-1 Created Stderr: Container roottestbackuprestoreoncluster-gw2-node3-1 Created Stderr: Container roottestbackuprestoreoncluster-gw2-node1-1 Created Stderr: Container roottestbackuprestoreoncluster-gw2-node2-1 Starting Stderr: Container roottestbackuprestoreoncluster-gw2-node3-1 Starting Stderr: Container roottestbackuprestoreoncluster-gw2-node1-1 Starting Stderr: Container roottestbackuprestoreoncluster-gw2-node3-1 Started Stderr: Container roottestbackuprestoreoncluster-gw2-node2-1 Started Stderr: Container roottestbackuprestoreoncluster-gw2-node1-1 Started ClickHouse instance created get_instance_ip instance_name=node1 http://localhost:None "GET /v1.46/containers/roottestbackuprestoreoncluster-gw2-node1-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node1 http://localhost:None "GET /v1.46/containers/roottestbackuprestoreoncluster-gw2-node1-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node1, ip: 172.16.5.5... http://localhost:None "GET /v1.46/containers/roottestbackuprestoreoncluster-gw2-node1-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/8377322328c200daa8d15336073ae4154dd0cabff93b6f8841a006eb3db5026f/json HTTP/1.1" 200 None Stderr: Container roottestconfigsubstitutions-gw6-zoo2-1 Running Stderr: Container roottestconfigsubstitutions-gw6-node1-1 Creating Stderr: Container roottestconfigsubstitutions-gw6-node5-1 Creating Stderr: Container roottestconfigsubstitutions-gw6-zoo3-1 Running Stderr: Container roottestconfigsubstitutions-gw6-node6-1 Creating Stderr: Container roottestconfigsubstitutions-gw6-node2-1 Creating Stderr: Container roottestconfigsubstitutions-gw6-node7-1 Creating Stderr: Container roottestconfigsubstitutions-gw6-node4-1 Creating Stderr: Container roottestconfigsubstitutions-gw6-node8-1 Creating Stderr: Container roottestconfigsubstitutions-gw6-zoo1-1 Running Stderr: Container roottestconfigsubstitutions-gw6-node3-1 Creating Stderr: Container roottestconfigsubstitutions-gw6-node5-1 Created Stderr: Container roottestconfigsubstitutions-gw6-node1-1 Created Stderr: Container roottestconfigsubstitutions-gw6-node6-1 Created Stderr: Container roottestconfigsubstitutions-gw6-node4-1 Created Stderr: Container roottestconfigsubstitutions-gw6-node7-1 Created Stderr: Container roottestconfigsubstitutions-gw6-node3-1 Created Stderr: Container roottestconfigsubstitutions-gw6-node8-1 Created Stderr: Container roottestconfigsubstitutions-gw6-node2-1 Created Stderr: Container roottestconfigsubstitutions-gw6-node5-1 Starting Stderr: Container roottestconfigsubstitutions-gw6-node1-1 Starting Stderr: Container roottestconfigsubstitutions-gw6-node6-1 Starting Stderr: Container roottestconfigsubstitutions-gw6-node7-1 Starting Stderr: Container roottestconfigsubstitutions-gw6-node2-1 Starting Stderr: Container roottestconfigsubstitutions-gw6-node8-1 Starting Stderr: Container roottestconfigsubstitutions-gw6-node3-1 Starting Stderr: Container roottestconfigsubstitutions-gw6-node4-1 Starting Stderr: Container roottestconfigsubstitutions-gw6-node3-1 Started Stderr: Container roottestconfigsubstitutions-gw6-node2-1 Started Stderr: Container roottestconfigsubstitutions-gw6-node1-1 Started Stderr: Container roottestconfigsubstitutions-gw6-node6-1 Started Stderr: Container roottestconfigsubstitutions-gw6-node5-1 Started Stderr: Container roottestconfigsubstitutions-gw6-node7-1 Started Stderr: Container roottestconfigsubstitutions-gw6-node8-1 Started Stderr: Container roottestconfigsubstitutions-gw6-node4-1 Started ClickHouse instance created get_instance_ip instance_name=node1 http://localhost:None "GET /v1.46/containers/roottestconfigsubstitutions-gw6-node1-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node1 http://localhost:None "GET /v1.46/containers/roottestconfigsubstitutions-gw6-node1-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node1, ip: 172.16.7.10... http://localhost:None "GET /v1.46/containers/roottestconfigsubstitutions-gw6-node1-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/d75db5277f501eb55ae8ab0da471553a3c017b1484f8785b9bd0ff8df8dbc089/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None Executing query ALTER TABLE test.tbl2 ATTACH PARTITION 197002 on node Executing query CREATE TABLE test.table_LocalClickHouse_complex_key_direct_ ( KeyField1 UInt64,KeyField2 String,UInt8_ UInt8,UInt16_ UInt16,UInt32_ UInt32,UInt64_ UInt64,Int8_ Int8,Int16_ Int16,Int32_ Int32,Int64_ Int64,UUID_ UUID,Date_ Date,DateTime_ DateTime,String_ String,Float32_ Float32,Float64_ Float64) ENGINE MergeTree ORDER BY tuple(); on local_node Executing query SELECT id, value FROM test_database_backup.test_table_2 on instance http://localhost:None "GET /v1.46/containers/8377322328c200daa8d15336073ae4154dd0cabff93b6f8841a006eb3db5026f/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/d75db5277f501eb55ae8ab0da471553a3c017b1484f8785b9bd0ff8df8dbc089/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None Executing query SELECT value FROM system.server_settings where name = 'allow_feature_tier' on instance http://localhost:None "GET /v1.46/containers/8377322328c200daa8d15336073ae4154dd0cabff93b6f8841a006eb3db5026f/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/d75db5277f501eb55ae8ab0da471553a3c017b1484f8785b9bd0ff8df8dbc089/json HTTP/1.1" 200 None Executing query INSERT INTO test.table_LocalClickHouse_complex_key_direct_ (KeyField1,KeyField2,UInt8_,UInt16_,UInt32_,UInt64_,Int8_,Int16_,Int32_,Int64_,UUID_,Date_,DateTime_,String_,Float32_,Float64_) values (1,'world',22,333,4444,55555,-6,-77,-888,-999,'550e8400-e29b-41d4-a716-446655440003','1973-06-28','1985-02-28 23:43:25','hello',22.543,3332154213.4),(2,'qwerty2',52,2345,6544,9191991,-2,-717,-81818,-92929,'550e8400-e29b-41d4-a716-446655440007','1975-09-28','2000-02-28 23:33:24','my',255.543,3332221.44) on local_node http://localhost:None "GET /v1.46/containers/8377322328c200daa8d15336073ae4154dd0cabff93b6f8841a006eb3db5026f/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/d75db5277f501eb55ae8ab0da471553a3c017b1484f8785b9bd0ff8df8dbc089/json HTTP/1.1" 200 None Executing query SELECT sleep(2) on node Executing query SELECT id, value FROM test_database_backup.test_table_3 on instance Executing query CREATE TABLE test_experimental (uid String, version UInt32, is_deleted UInt8) ENGINE = ReplacingMergeTree(version, is_deleted) ORDER by (uid) SETTINGS allow_experimental_replacing_merge_with_cleanup=1; on instance http://localhost:None "GET /v1.46/containers/8377322328c200daa8d15336073ae4154dd0cabff93b6f8841a006eb3db5026f/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/d75db5277f501eb55ae8ab0da471553a3c017b1484f8785b9bd0ff8df8dbc089/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/8377322328c200daa8d15336073ae4154dd0cabff93b6f8841a006eb3db5026f/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/d75db5277f501eb55ae8ab0da471553a3c017b1484f8785b9bd0ff8df8dbc089/json HTTP/1.1" 200 None Executing query CREATE DATABASE IF NOT EXISTS test on local_node http://localhost:None "GET /v1.46/containers/8377322328c200daa8d15336073ae4154dd0cabff93b6f8841a006eb3db5026f/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/d75db5277f501eb55ae8ab0da471553a3c017b1484f8785b9bd0ff8df8dbc089/json HTTP/1.1" 200 None run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:True cmd: ['bash', '-c', 'ps -C clickhouse'] Command:[docker exec -u root roottestdatabasebackup-gw1-instance-1 bash -c ps -C clickhouse] http://localhost:None "GET /v1.46/containers/8377322328c200daa8d15336073ae4154dd0cabff93b6f8841a006eb3db5026f/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/d75db5277f501eb55ae8ab0da471553a3c017b1484f8785b9bd0ff8df8dbc089/json HTTP/1.1" 200 None Executing query select count() from system.tables where database = 'system' and name in ['query_log', 'text_log'] on node2 Executing query CREATE TABLE test.table_LocalClickHouse_range_hashed_ ( KeyField1 UInt64,KeyField2 Date,UInt8_ UInt8,UInt16_ UInt16,UInt32_ UInt32,UInt64_ UInt64,Int8_ Int8,Int16_ Int16,Int32_ Int32,Int64_ Int64,UUID_ UUID,Date_ Date,DateTime_ DateTime,String_ String,Float32_ Float32,Float64_ Float64,StartDate Date,EndDate Date) ENGINE MergeTree ORDER BY tuple(); on local_node Stdout: PID TTY TIME CMD Stdout: 8 ? 00:00:02 clickhouse run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', 'pkill clickhouse'] Command:[docker exec -u root roottestdatabasebackup-gw1-instance-1 bash -c pkill clickhouse] http://localhost:None "GET /v1.46/containers/8377322328c200daa8d15336073ae4154dd0cabff93b6f8841a006eb3db5026f/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/d75db5277f501eb55ae8ab0da471553a3c017b1484f8785b9bd0ff8df8dbc089/json HTTP/1.1" 200 None run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] http://localhost:None "GET /v1.46/containers/8377322328c200daa8d15336073ae4154dd0cabff93b6f8841a006eb3db5026f/json HTTP/1.1" 200 None run container_id:roottestallowfeaturetier-gw8-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "sed -i 's/1/0/g' /etc/clickhouse-server/config.d/allow_feature_tier.xml"] Command:[docker exec roottestallowfeaturetier-gw8-instance-1 bash -c sed -i 's/1/0/g' /etc/clickhouse-server/config.d/allow_feature_tier.xml] http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/d75db5277f501eb55ae8ab0da471553a3c017b1484f8785b9bd0ff8df8dbc089/json HTTP/1.1" 200 None Stdout:8 Executing query SYSTEM RELOAD CONFIG on instance http://localhost:None "GET /v1.46/containers/8377322328c200daa8d15336073ae4154dd0cabff93b6f8841a006eb3db5026f/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/d75db5277f501eb55ae8ab0da471553a3c017b1484f8785b9bd0ff8df8dbc089/json HTTP/1.1" 200 None Executing query INSERT INTO test.table_LocalClickHouse_range_hashed_ (KeyField1,StartDate,EndDate,KeyField2,UInt8_,UInt16_,UInt32_,UInt64_,Int8_,Int16_,Int32_,Int64_,UUID_,Date_,DateTime_,String_,Float32_,Float64_) values (1,'2019-02-01','2019-02-28','2019-02-10',22,333,4444,55555,-6,-77,-888,-999,'550e8400-e29b-41d4-a716-446655440003','1973-06-28','1985-02-28 23:43:25','hello',22.543,3332154213.4),(2,'2019-04-01','2019-04-28','2019-04-10',11,3223,41444,52515,-65,-747,-8388,-9099,'550e8400-e29b-41d4-a716-446655440004','1973-06-29','2002-02-28 23:23:25','!!!!',32.543,3332543.4) on local_node Executing query rename table system.text_log to system.text_log_3_test on node2 http://localhost:None "GET /v1.46/containers/8377322328c200daa8d15336073ae4154dd0cabff93b6f8841a006eb3db5026f/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/d75db5277f501eb55ae8ab0da471553a3c017b1484f8785b9bd0ff8df8dbc089/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/8377322328c200daa8d15336073ae4154dd0cabff93b6f8841a006eb3db5026f/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/d75db5277f501eb55ae8ab0da471553a3c017b1484f8785b9bd0ff8df8dbc089/json HTTP/1.1" 200 None Executing query SELECT value FROM system.server_settings where name = 'allow_feature_tier' on instance http://localhost:None "GET /v1.46/containers/8377322328c200daa8d15336073ae4154dd0cabff93b6f8841a006eb3db5026f/json HTTP/1.1" 200 None Executing query rename table system.query_log to system.query_log_3_test on node2 http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/d75db5277f501eb55ae8ab0da471553a3c017b1484f8785b9bd0ff8df8dbc089/json HTTP/1.1" 200 None Executing query system reload dictionaries on local_node http://localhost:None "GET /v1.46/containers/8377322328c200daa8d15336073ae4154dd0cabff93b6f8841a006eb3db5026f/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/d75db5277f501eb55ae8ab0da471553a3c017b1484f8785b9bd0ff8df8dbc089/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/8377322328c200daa8d15336073ae4154dd0cabff93b6f8841a006eb3db5026f/json HTTP/1.1" 200 None Executing query system flush logs on node2 http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None Executing query CREATE TABLE test_experimental (uid String, version UInt32, is_deleted UInt8) ENGINE = ReplacingMergeTree(version, is_deleted) ORDER by (uid) SETTINGS allow_experimental_replacing_merge_with_cleanup=1; on instance http://localhost:None "GET /v1.46/containers/d75db5277f501eb55ae8ab0da471553a3c017b1484f8785b9bd0ff8df8dbc089/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/8377322328c200daa8d15336073ae4154dd0cabff93b6f8841a006eb3db5026f/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/d75db5277f501eb55ae8ab0da471553a3c017b1484f8785b9bd0ff8df8dbc089/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/8377322328c200daa8d15336073ae4154dd0cabff93b6f8841a006eb3db5026f/json HTTP/1.1" 200 None Executing query select dictGet('LocalClickHouse_complex_key_cache_', 'UInt8_', tuple(toUInt64(1),toString('world'))) on local_node http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/d75db5277f501eb55ae8ab0da471553a3c017b1484f8785b9bd0ff8df8dbc089/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/8377322328c200daa8d15336073ae4154dd0cabff93b6f8841a006eb3db5026f/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None Executing query SELECT engine_full FROM system.tables WHERE name = 'test_experimental' on instance http://localhost:None "GET /v1.46/containers/d75db5277f501eb55ae8ab0da471553a3c017b1484f8785b9bd0ff8df8dbc089/json HTTP/1.1" 200 None run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] http://localhost:None "GET /v1.46/containers/8377322328c200daa8d15336073ae4154dd0cabff93b6f8841a006eb3db5026f/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/d75db5277f501eb55ae8ab0da471553a3c017b1484f8785b9bd0ff8df8dbc089/json HTTP/1.1" 200 None Stdout:8 http://localhost:None "GET /v1.46/containers/8377322328c200daa8d15336073ae4154dd0cabff93b6f8841a006eb3db5026f/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/d75db5277f501eb55ae8ab0da471553a3c017b1484f8785b9bd0ff8df8dbc089/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/8377322328c200daa8d15336073ae4154dd0cabff93b6f8841a006eb3db5026f/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/d75db5277f501eb55ae8ab0da471553a3c017b1484f8785b9bd0ff8df8dbc089/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/8377322328c200daa8d15336073ae4154dd0cabff93b6f8841a006eb3db5026f/json HTTP/1.1" 200 None run container_id:roottestallowfeaturetier-gw8-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "sed -i 's/0/1/g' /etc/clickhouse-server/config.d/allow_feature_tier.xml"] Command:[docker exec roottestallowfeaturetier-gw8-instance-1 bash -c sed -i 's/0/1/g' /etc/clickhouse-server/config.d/allow_feature_tier.xml] http://localhost:None "GET /v1.46/containers/d75db5277f501eb55ae8ab0da471553a3c017b1484f8785b9bd0ff8df8dbc089/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/8377322328c200daa8d15336073ae4154dd0cabff93b6f8841a006eb3db5026f/json HTTP/1.1" 200 None Executing query SYSTEM RELOAD CONFIG on instance http://localhost:None "GET /v1.46/containers/d75db5277f501eb55ae8ab0da471553a3c017b1484f8785b9bd0ff8df8dbc089/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/8377322328c200daa8d15336073ae4154dd0cabff93b6f8841a006eb3db5026f/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/d75db5277f501eb55ae8ab0da471553a3c017b1484f8785b9bd0ff8df8dbc089/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None Executing query SELECT * FROM test.tbl2 ORDER BY p on node Executing query select dictGetUInt8('LocalClickHouse_complex_key_cache_', 'UInt8_', tuple(toUInt64(1),toString('world'))) on local_node http://localhost:None "GET /v1.46/containers/8377322328c200daa8d15336073ae4154dd0cabff93b6f8841a006eb3db5026f/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/d75db5277f501eb55ae8ab0da471553a3c017b1484f8785b9bd0ff8df8dbc089/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/8377322328c200daa8d15336073ae4154dd0cabff93b6f8841a006eb3db5026f/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/d75db5277f501eb55ae8ab0da471553a3c017b1484f8785b9bd0ff8df8dbc089/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/8377322328c200daa8d15336073ae4154dd0cabff93b6f8841a006eb3db5026f/json HTTP/1.1" 200 None Executing query SELECT value FROM system.server_settings where name = 'allow_feature_tier' on instance Stderr: mysql80 Pulling Stderr: node Pulling Stderr: cea172a6e83b Pulling fs layer Stderr: 28e01aa53f13 Pulling fs layer Stderr: e5fa3211d7a7 Pulling fs layer Stderr: 753b8441f7e6 Pulling fs layer Stderr: b1339a14fa1a Pulling fs layer Stderr: be386ff914e3 Pulling fs layer http://localhost:None "GET /v1.46/containers/d75db5277f501eb55ae8ab0da471553a3c017b1484f8785b9bd0ff8df8dbc089/json HTTP/1.1" 200 None Stderr: 93272c957f26 Pulling fs layer Stderr: c106a4902288 Pulling fs layer Stderr: 036f4325df2d Pulling fs layer Stderr: 1d34979e7120 Pulling fs layer Stderr: de67a2f637e5 Pulling fs layer Stderr: 753b8441f7e6 Waiting Stderr: b1339a14fa1a Waiting Stderr: 93272c957f26 Waiting Stderr: be386ff914e3 Waiting Stderr: c106a4902288 Waiting Stderr: de67a2f637e5 Waiting Stderr: 1d34979e7120 Waiting Stderr: 036f4325df2d Waiting Stderr: cea172a6e83b Downloading [> ] 509.6kB/49.09MB Stderr: e5fa3211d7a7 Downloading [> ] 13.78kB/983kB Stderr: 28e01aa53f13 Downloading [==================================================>] 882B/882B Stderr: 28e01aa53f13 Download complete Stderr: e5fa3211d7a7 Downloading [==================================================>] 983kB/983kB Stderr: e5fa3211d7a7 Download complete Stderr: 753b8441f7e6 Downloading [> ] 75.51kB/6.897MB Stderr: b1339a14fa1a Downloading [==================================================>] 2.606kB/2.606kB Stderr: b1339a14fa1a Verifying Checksum Stderr: b1339a14fa1a Download complete Stderr: be386ff914e3 Downloading [==================================================>] 333B/333B Stderr: be386ff914e3 Verifying Checksum Stderr: 93272c957f26 Downloading [> ] 503.2kB/49.62MB Stderr: 753b8441f7e6 Download complete Stderr: c106a4902288 Downloading [==================================================>] 315B/315B Stderr: c106a4902288 Download complete Stderr: 036f4325df2d Downloading [> ] 524.8kB/125.3MB Stderr: cea172a6e83b Downloading [===========================================> ] 42.33MB/49.09MB Stderr: cea172a6e83b Verifying Checksum Stderr: cea172a6e83b Download complete Stderr: 1d34979e7120 Downloading [==================================> ] 3.647kB/5.327kB Stderr: 1d34979e7120 Downloading [==================================================>] 5.327kB/5.327kB Stderr: 1d34979e7120 Download complete Stderr: 93272c957f26 Downloading [========================================> ] 40.27MB/49.62MB Stderr: de67a2f637e5 Downloading [==================================================>] 122B/122B Stderr: de67a2f637e5 Verifying Checksum Stderr: de67a2f637e5 Download complete Stderr: 93272c957f26 Verifying Checksum Stderr: 93272c957f26 Download complete Stderr: 036f4325df2d Downloading [===============> ] 38.34MB/125.3MB Stderr: cea172a6e83b Extracting [> ] 491.5kB/49.09MB Stderr: 036f4325df2d Downloading [===============================> ] 79.01MB/125.3MB Stderr: cea172a6e83b Extracting [=> ] 1.475MB/49.09MB Stderr: node Pulled Stderr: 036f4325df2d Downloading [===============================================> ] 118MB/125.3MB Stderr: 036f4325df2d Verifying Checksum Stderr: 036f4325df2d Download complete Stderr: cea172a6e83b Extracting [=======> ] 6.881MB/49.09MB Stderr: cea172a6e83b Extracting [=======> ] 7.373MB/49.09MB Stderr: cea172a6e83b Extracting [===============> ] 14.75MB/49.09MB Stderr: cea172a6e83b Extracting [=======================> ] 23.1MB/49.09MB Stderr: cea172a6e83b Extracting [==============================> ] 29.98MB/49.09MB Stderr: cea172a6e83b Extracting [=================================> ] 32.44MB/49.09MB Stderr: cea172a6e83b Extracting [============================================> ] 43.25MB/49.09MB Stderr: cea172a6e83b Extracting [==============================================> ] 45.71MB/49.09MB Stderr: cea172a6e83b Extracting [===============================================> ] 46.2MB/49.09MB Stderr: cea172a6e83b Extracting [==================================================>] 49.09MB/49.09MB Stderr: cea172a6e83b Pull complete Stderr: 28e01aa53f13 Extracting [==================================================>] 882B/882B Stderr: 28e01aa53f13 Extracting [==================================================>] 882B/882B Stderr: 28e01aa53f13 Pull complete Stderr: e5fa3211d7a7 Extracting [=> ] 32.77kB/983kB Stderr: e5fa3211d7a7 Extracting [==================================================>] 983kB/983kB Stderr: e5fa3211d7a7 Extracting [==================================================>] 983kB/983kB Stderr: e5fa3211d7a7 Pull complete Stderr: 753b8441f7e6 Extracting [> ] 98.3kB/6.897MB Stderr: 753b8441f7e6 Extracting [=================> ] 2.359MB/6.897MB Stderr: 753b8441f7e6 Extracting [==========================> ] 3.637MB/6.897MB Stderr: 753b8441f7e6 Extracting [==================================================>] 6.897MB/6.897MB Stderr: 753b8441f7e6 Pull complete Stderr: b1339a14fa1a Extracting [==================================================>] 2.606kB/2.606kB Stderr: b1339a14fa1a Extracting [==================================================>] 2.606kB/2.606kB Stderr: b1339a14fa1a Pull complete Stderr: be386ff914e3 Extracting [==================================================>] 333B/333B Stderr: be386ff914e3 Extracting [==================================================>] 333B/333B Stderr: be386ff914e3 Pull complete Stderr: 93272c957f26 Extracting [> ] 524.3kB/49.62MB Stderr: 93272c957f26 Extracting [=====> ] 5.243MB/49.62MB Stderr: 93272c957f26 Extracting [=========> ] 9.437MB/49.62MB Stderr: 93272c957f26 Extracting [=============> ] 13.63MB/49.62MB Stderr: 93272c957f26 Extracting [================> ] 16.25MB/49.62MB Stderr: 93272c957f26 Extracting [=====================> ] 21.5MB/49.62MB Stderr: 93272c957f26 Extracting [===========================> ] 27.26MB/49.62MB Stderr: 93272c957f26 Extracting [=================================> ] 33.03MB/49.62MB Stderr: 93272c957f26 Extracting [======================================> ] 38.27MB/49.62MB Stderr: 93272c957f26 Extracting [===========================================> ] 43.52MB/49.62MB Stderr: 93272c957f26 Extracting [================================================> ] 47.71MB/49.62MB Stderr: 93272c957f26 Extracting [==================================================>] 49.62MB/49.62MB Stderr: 93272c957f26 Pull complete Stderr: c106a4902288 Extracting [==================================================>] 315B/315B Stderr: c106a4902288 Extracting [==================================================>] 315B/315B Stderr: c106a4902288 Pull complete Stderr: 036f4325df2d Extracting [> ] 557.1kB/125.3MB Stderr: 036f4325df2d Extracting [==> ] 5.571MB/125.3MB Stderr: 036f4325df2d Extracting [====> ] 12.26MB/125.3MB Stderr: 036f4325df2d Extracting [======> ] 17.27MB/125.3MB Stderr: 036f4325df2d Extracting [========> ] 21.73MB/125.3MB Stderr: 036f4325df2d Extracting [=========> ] 24.51MB/125.3MB Stderr: 036f4325df2d Extracting [===========> ] 28.41MB/125.3MB Stderr: 036f4325df2d Extracting [============> ] 31.2MB/125.3MB Stderr: 036f4325df2d Extracting [============> ] 31.75MB/125.3MB Stderr: 036f4325df2d Extracting [============> ] 32.31MB/125.3MB Stderr: 036f4325df2d Extracting [=============> ] 32.87MB/125.3MB Stderr: 036f4325df2d Extracting [=============> ] 33.42MB/125.3MB Stderr: 036f4325df2d Extracting [=============> ] 33.98MB/125.3MB Stderr: 036f4325df2d Extracting [=============> ] 34.54MB/125.3MB Stderr: 036f4325df2d Extracting [==============> ] 35.09MB/125.3MB Stderr: 036f4325df2d Extracting [==============> ] 35.65MB/125.3MB Stderr: 036f4325df2d Extracting [==============> ] 36.21MB/125.3MB Stderr: 036f4325df2d Extracting [==============> ] 36.77MB/125.3MB Stderr: 036f4325df2d Extracting [==============> ] 37.32MB/125.3MB Stderr: 036f4325df2d Extracting [===============> ] 37.88MB/125.3MB Stderr: 036f4325df2d Extracting [===============> ] 38.44MB/125.3MB Stderr: 036f4325df2d Extracting [===============> ] 38.99MB/125.3MB Stderr: 036f4325df2d Extracting [===============> ] 39.55MB/125.3MB Stderr: 036f4325df2d Extracting [================> ] 40.11MB/125.3MB Stderr: 036f4325df2d Extracting [================> ] 40.67MB/125.3MB Stderr: 036f4325df2d Extracting [================> ] 41.22MB/125.3MB Stderr: 036f4325df2d Extracting [================> ] 41.78MB/125.3MB Stderr: 036f4325df2d Extracting [================> ] 42.34MB/125.3MB Stderr: 036f4325df2d Extracting [=================> ] 42.89MB/125.3MB Stderr: 036f4325df2d Extracting [=================> ] 43.45MB/125.3MB Stderr: 036f4325df2d Extracting [==================> ] 45.68MB/125.3MB Stderr: 036f4325df2d Extracting [==================> ] 46.24MB/125.3MB Stderr: 036f4325df2d Extracting [==================> ] 46.79MB/125.3MB Stderr: 036f4325df2d Extracting [===================> ] 48.46MB/125.3MB Stderr: 036f4325df2d Extracting [====================> ] 51.25MB/125.3MB Stderr: 036f4325df2d Extracting [=======================> ] 57.93MB/125.3MB Stderr: 036f4325df2d Extracting [=========================> ] 63.5MB/125.3MB Stderr: 036f4325df2d Extracting [==========================> ] 66.29MB/125.3MB Stderr: 036f4325df2d Extracting [============================> ] 70.19MB/125.3MB Stderr: 036f4325df2d Extracting [==============================> ] 75.2MB/125.3MB Stderr: 036f4325df2d Extracting [================================> ] 80.77MB/125.3MB Stderr: 036f4325df2d Extracting [=================================> ] 84.67MB/125.3MB Stderr: 036f4325df2d Extracting [====================================> ] 90.24MB/125.3MB Stderr: 036f4325df2d Extracting [=====================================> ] 94.14MB/125.3MB Stderr: 036f4325df2d Extracting [=======================================> ] 99.71MB/125.3MB Stderr: 036f4325df2d Extracting [========================================> ] 100.3MB/125.3MB Stderr: 036f4325df2d Extracting [=========================================> ] 104.2MB/125.3MB Stderr: 036f4325df2d Extracting [===========================================> ] 109.2MB/125.3MB Stderr: 036f4325df2d Extracting [============================================> ] 112.5MB/125.3MB Stderr: 036f4325df2d Extracting [==============================================> ] 115.3MB/125.3MB Stderr: 036f4325df2d Extracting [===============================================> ] 118.7MB/125.3MB Stderr: 036f4325df2d Extracting [================================================> ] 120.9MB/125.3MB Stderr: 036f4325df2d Extracting [=================================================> ] 124.2MB/125.3MB Stderr: 036f4325df2d Extracting [==================================================>] 125.3MB/125.3MB Stderr: 036f4325df2d Pull complete Stderr: 1d34979e7120 Extracting [==================================================>] 5.327kB/5.327kB Stderr: 1d34979e7120 Extracting [==================================================>] 5.327kB/5.327kB Stderr: 1d34979e7120 Pull complete Stderr: de67a2f637e5 Extracting [==================================================>] 122B/122B Stderr: de67a2f637e5 Extracting [==================================================>] 122B/122B Stderr: de67a2f637e5 Pull complete Stderr: mysql80 Pulled Setup MySQL 8 Command:[docker compose --project-name roottestdictionariesalllayoutsseparatesourcesmysql-gw3 --env-file /ClickHouse/tests/integration/test_dictionaries_all_layouts_separate_sources/_instances-mysql-0-gw3/.env --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_mysql_8_0.yml --verbose up -d] http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None Executing query ALTER TABLE test.tbl2 UPDATE k=10 WHERE 1 on node http://localhost:None "GET /v1.46/containers/8377322328c200daa8d15336073ae4154dd0cabff93b6f8841a006eb3db5026f/json HTTP/1.1" 200 None Executing query select dictHas('LocalClickHouse_complex_key_cache_', tuple(toUInt64(1),toString('world'))) on local_node http://localhost:None "GET /v1.46/containers/d75db5277f501eb55ae8ab0da471553a3c017b1484f8785b9bd0ff8df8dbc089/json HTTP/1.1" 200 None ClickHouse node1 started get_instance_ip instance_name=node2 http://localhost:None "GET /v1.46/containers/roottestconfigsubstitutions-gw6-node2-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node2 http://localhost:None "GET /v1.46/containers/roottestconfigsubstitutions-gw6-node2-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node2, ip: 172.16.7.9... http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottestconfigsubstitutions-gw6-node2-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/3513206d35403c32e1b5d524abf2929813739c9014805b4542f22f894d1fc957/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/8377322328c200daa8d15336073ae4154dd0cabff93b6f8841a006eb3db5026f/json HTTP/1.1" 200 None Executing query select count() from system.tables where database = 'system' and name in ['query_log', 'text_log'] on node2 run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/3513206d35403c32e1b5d524abf2929813739c9014805b4542f22f894d1fc957/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/8377322328c200daa8d15336073ae4154dd0cabff93b6f8841a006eb3db5026f/json HTTP/1.1" 200 None run container_id:roottestallowfeaturetier-gw8-instance-1 detach:False nothrow:True cmd: ['bash', '-c', 'ps -C clickhouse'] Command:[docker exec -u root roottestallowfeaturetier-gw8-instance-1 bash -c ps -C clickhouse] Stdout:8 Executing query SELECT sleep(2) on node Stdout: PID TTY TIME CMD Stdout: 8 ? 00:00:05 clickhouse run container_id:roottestallowfeaturetier-gw8-instance-1 detach:False nothrow:False cmd: ['bash', '-c', 'pkill clickhouse'] Command:[docker exec -u root roottestallowfeaturetier-gw8-instance-1 bash -c pkill clickhouse] http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/3513206d35403c32e1b5d524abf2929813739c9014805b4542f22f894d1fc957/json HTTP/1.1" 200 None ClickHouse node2 started get_instance_ip instance_name=node3 http://localhost:None "GET /v1.46/containers/roottestconfigsubstitutions-gw6-node3-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node3 http://localhost:None "GET /v1.46/containers/8377322328c200daa8d15336073ae4154dd0cabff93b6f8841a006eb3db5026f/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottestconfigsubstitutions-gw6-node3-1/json HTTP/1.1" 200 None run container_id:roottestallowfeaturetier-gw8-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestallowfeaturetier-gw8-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Waiting for ClickHouse start in node3, ip: 172.16.7.6... Executing query select dictGetOrDefault('LocalClickHouse_complex_key_cache_', 'UInt8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt8(55)) on local_node http://localhost:None "GET /v1.46/containers/roottestconfigsubstitutions-gw6-node3-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/ff339805d5b7adb039a64a4953ae93e3ee6ceae1ebad2cb61142859393e97124/json HTTP/1.1" 200 None ClickHouse node3 started get_instance_ip instance_name=node4 http://localhost:None "GET /v1.46/containers/roottestconfigsubstitutions-gw6-node4-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node4 http://localhost:None "GET /v1.46/containers/roottestconfigsubstitutions-gw6-node4-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node4, ip: 172.16.7.12... http://localhost:None "GET /v1.46/containers/roottestconfigsubstitutions-gw6-node4-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/9303f6b4703432d7dac8258a10bf1b73383ebece7607c222587344053e237a6f/json HTTP/1.1" 200 None ClickHouse node4 started get_instance_ip instance_name=node5 http://localhost:None "GET /v1.46/containers/roottestconfigsubstitutions-gw6-node5-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node5 http://localhost:None "GET /v1.46/containers/roottestconfigsubstitutions-gw6-node5-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node5, ip: 172.16.7.5... http://localhost:None "GET /v1.46/containers/roottestconfigsubstitutions-gw6-node5-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/39936e05f57d15c8ac09fdd8445ffd347784af08c807efed1b5559df0e444b09/json HTTP/1.1" 200 None ClickHouse node5 started get_instance_ip instance_name=node6 http://localhost:None "GET /v1.46/containers/roottestconfigsubstitutions-gw6-node6-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node6 http://localhost:None "GET /v1.46/containers/roottestconfigsubstitutions-gw6-node6-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node6, ip: 172.16.7.11... http://localhost:None "GET /v1.46/containers/roottestconfigsubstitutions-gw6-node6-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/c27d4632f548d975abce93edbd58f57c6b137a541855887cb693d676c9def1b2/json HTTP/1.1" 200 None ClickHouse node6 started get_instance_ip instance_name=node7 http://localhost:None "GET /v1.46/containers/roottestconfigsubstitutions-gw6-node7-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node7 http://localhost:None "GET /v1.46/containers/roottestconfigsubstitutions-gw6-node7-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node7, ip: 172.16.7.8... http://localhost:None "GET /v1.46/containers/roottestconfigsubstitutions-gw6-node7-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/3aa7e5e396a6179e41ebe1b14bf6afcd27d5645137c238bd30c84238089bb606/json HTTP/1.1" 200 None Stdout:8 http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/8377322328c200daa8d15336073ae4154dd0cabff93b6f8841a006eb3db5026f/json HTTP/1.1" 200 None ClickHouse node1 started get_instance_ip instance_name=node2 http://localhost:None "GET /v1.46/containers/roottestbackuprestoreoncluster-gw2-node2-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node2 http://localhost:None "GET /v1.46/containers/roottestbackuprestoreoncluster-gw2-node2-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node2, ip: 172.16.5.6... http://localhost:None "GET /v1.46/containers/roottestbackuprestoreoncluster-gw2-node2-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/ad2b3876200af3e3c2d76cc86c4d85d24060856fd73a24c890901dac7aaaee14/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/3aa7e5e396a6179e41ebe1b14bf6afcd27d5645137c238bd30c84238089bb606/json HTTP/1.1" 200 None ClickHouse node7 started get_instance_ip instance_name=node8 http://localhost:None "GET /v1.46/containers/roottestconfigsubstitutions-gw6-node8-1/json HTTP/1.1" 200 None Executing query rename table system.text_log to system.text_log_4_test on node2 get_instance_ip instance_name=node8 http://localhost:None "GET /v1.46/containers/roottestconfigsubstitutions-gw6-node8-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node8, ip: 172.16.7.7... http://localhost:None "GET /v1.46/containers/roottestconfigsubstitutions-gw6-node8-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/040acaa439d305ce37e34ecb613cb5b9d5cab7d7883953b74148c54201c1ac6f/json HTTP/1.1" 200 None ClickHouse node8 started Executing query DROP DATABASE IF EXISTS db1 SYNC on node5 http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/ad2b3876200af3e3c2d76cc86c4d85d24060856fd73a24c890901dac7aaaee14/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/ad2b3876200af3e3c2d76cc86c4d85d24060856fd73a24c890901dac7aaaee14/json HTTP/1.1" 200 None ClickHouse node2 started get_instance_ip instance_name=node3 http://localhost:None "GET /v1.46/containers/roottestbackuprestoreoncluster-gw2-node3-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node3 http://localhost:None "GET /v1.46/containers/roottestbackuprestoreoncluster-gw2-node3-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node3, ip: 172.16.5.7... http://localhost:None "GET /v1.46/containers/roottestbackuprestoreoncluster-gw2-node3-1/json HTTP/1.1" 200 None Executing query select dictGetUInt8OrDefault('LocalClickHouse_complex_key_cache_', 'UInt8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt8(55)) on local_node http://localhost:None "GET /v1.46/containers/fdd0f5d75636a6e392341a62144913161c851e21e6b457f8d60f99d48caae9f5/json HTTP/1.1" 200 None ClickHouse node3 started Executing query CREATE TABLE tbl ON CLUSTER 'cluster' (x UInt8) ENGINE=ReplicatedMergeTree('/clickhouse/tables/tbl/', '{replica}')ORDER BY x on node1 Stderr:time="2025-04-02T02:18:14Z" level=trace msg="Docker Desktop integration not enabled" Stderr: Network roottestdictionariesalllayoutsseparatesourcesmysql-gw3_default Creating Stderr: Network roottestdictionariesalllayoutsseparatesourcesmysql-gw3_default Created Stderr: Container roottestdictionariesalllayoutsseparatesourcesmysql-gw3-mysql80-1 Creating Stderr: Container roottestdictionariesalllayoutsseparatesourcesmysql-gw3-mysql80-1 Created Stderr: Container roottestdictionariesalllayoutsseparatesourcesmysql-gw3-mysql80-1 Starting Stderr: Container roottestdictionariesalllayoutsseparatesourcesmysql-gw3-mysql80-1 Started Stderr:time="2025-04-02T02:18:14Z" level=debug msg="otel error" error="" Stderr:time="2025-04-02T02:18:14Z" level=debug msg="otel error" error="" get_instance_ip instance_name=mysql80 http://localhost:None "GET /v1.46/containers/roottestdictionariesalllayoutsseparatesourcesmysql-gw3-mysql80-1/json HTTP/1.1" 200 None Can't connect to MySQL 8 (2003, "Can't connect to MySQL server on '172.16.10.2' ([Errno 111] Connection refused)") http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None Executing query rename table system.query_log to system.query_log_4_test on node2 Executing query CREATE DATABASE db1 on node5 http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None Executing query select dictGet('LocalClickHouse_complex_key_cache_', 'UInt16_', tuple(toUInt64(1),toString('world'))) on local_node http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None Executing query system flush logs on node2 Executing query CREATE TABLE db1.test_table(date Date, k1 String, v1 Int32) ENGINE = MergeTree(date, (k1, date), 8192) on node5 http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None Executing query INSERT INTO tbl VALUES (1) on node1 http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None Can't connect to MySQL 8 (2003, "Can't connect to MySQL server on '172.16.10.2' ([Errno 111] Connection refused)") Executing query INSERT INTO db1.test_table VALUES('2000-01-01', 'test_key', 1) on node5 Executing query select dictGetUInt16('LocalClickHouse_complex_key_cache_', 'UInt16_', tuple(toUInt64(1),toString('world'))) on local_node http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None Executing query BACKUP TABLE tbl TO Disk('backups', '1') ASYNC on node1 via HTTP interface Starting new HTTP connection (1): 172.16.5.5:8123 Stdout:8 http://172.16.5.5:8123 "GET /?query=BACKUP+TABLE+tbl++TO+Disk%28%27backups%27%2C+%271%27%29+ASYNC HTTP/1.1" 200 None Executing query BACKUP TABLE tbl TO Disk('backups', '1') ASYNC on node2 via HTTP interface Starting new HTTP connection (1): 172.16.5.6:8123 http://172.16.5.6:8123 "GET /?query=BACKUP+TABLE+tbl++TO+Disk%28%27backups%27%2C+%271%27%29+ASYNC HTTP/1.1" 200 None Executing query SELECT status FROM system.backups WHERE id IN ['fe5dfd90-241a-4afe-864e-d3f2742a4a8b', 'e0325bef-595b-4312-a9a2-bbadb7f61ba0'] AND status == 'CREATING_BACKUP' on node1 run container_id:roottestallowfeaturetier-gw8-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestallowfeaturetier-gw8-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None Stdout:8 Executing query SELECT name FROM system.databases WHERE name = 'db1' on node5 Executing query select dictHas('LocalClickHouse_complex_key_cache_', tuple(toUInt64(1),toString('world'))) on local_node http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None Can't connect to MySQL 8 (2003, "Can't connect to MySQL server on '172.16.10.2' ([Errno 111] Connection refused)") Executing query SELECT status FROM system.backups WHERE id IN ['fe5dfd90-241a-4afe-864e-d3f2742a4a8b', 'e0325bef-595b-4312-a9a2-bbadb7f61ba0'] AND status == 'CREATING_BACKUP' on node2 ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_dictionaries_redis/_instances-0-gw0/.env --project-name roottestdictionariesredis-gw0 --file /ClickHouse/tests/integration/test_dictionaries_redis/_instances-0-gw0/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_redis.yml up -d --no-recreate') Command:[docker compose --env-file /ClickHouse/tests/integration/test_dictionaries_redis/_instances-0-gw0/.env --project-name roottestdictionariesredis-gw0 --file /ClickHouse/tests/integration/test_dictionaries_redis/_instances-0-gw0/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_redis.yml up -d --no-recreate] http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None Executing query select dictGetOrDefault('LocalClickHouse_complex_key_cache_', 'UInt16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt16(66)) on local_node Executing query SELECT name FROM system.tables WHERE database = 'db1' AND name = 'test_table' on node5 http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None Executing query select count() from system.tables where database = 'system' and name in ['query_log', 'text_log'] on node2 http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None Executing query SELECT count() FROM system.backups WHERE id IN ['fe5dfd90-241a-4afe-864e-d3f2742a4a8b', 'e0325bef-595b-4312-a9a2-bbadb7f61ba0'] AND status == 'BACKUP_CREATED' on node1 Stderr: Container roottestdictionariesredis-gw0-redis1-1 Running Executing query SELECT name FROM system.columns WHERE database = 'db1' AND table = 'test_table' on node5 Stderr: Container roottestdictionariesredis-gw0-node-1 Creating Stderr: Container roottestdictionariesredis-gw0-node-1 Created Stderr: Container roottestdictionariesredis-gw0-node-1 Starting Stderr: Container roottestdictionariesredis-gw0-node-1 Started ClickHouse instance created get_instance_ip instance_name=node http://localhost:None "GET /v1.46/containers/roottestdictionariesredis-gw0-node-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node http://localhost:None "GET /v1.46/containers/roottestdictionariesredis-gw0-node-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node, ip: 172.16.9.3... http://localhost:None "GET /v1.46/containers/roottestdictionariesredis-gw0-node-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/d86a3a10bee16809ac83c2260bef25842cac0f676a7c8b3c2e694f9fbc6490f8/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None Executing query select dictGetUInt16OrDefault('LocalClickHouse_complex_key_cache_', 'UInt16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt16(66)) on local_node http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/d86a3a10bee16809ac83c2260bef25842cac0f676a7c8b3c2e694f9fbc6490f8/json HTTP/1.1" 200 None Can't connect to MySQL 8 (2003, "Can't connect to MySQL server on '172.16.10.2' ([Errno 111] Connection refused)") Executing query rename table system.text_log to system.text_log_5_test on node2 http://localhost:None "GET /v1.46/containers/b04d8405d0aa47276e115f5f41699a99a0a0b080c3a3818f94ccf4994f28d29b/json HTTP/1.1" 200 None ClickHouse node1 started get_instance_ip instance_name=node2 http://localhost:None "GET /v1.46/containers/d86a3a10bee16809ac83c2260bef25842cac0f676a7c8b3c2e694f9fbc6490f8/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/roottestclusterallreplicas-gw4-node2-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node2 http://localhost:None "GET /v1.46/containers/roottestclusterallreplicas-gw4-node2-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node2, ip: 172.16.6.2... http://localhost:None "GET /v1.46/containers/roottestclusterallreplicas-gw4-node2-1/json HTTP/1.1" 200 None Executing query SELECT count() FROM system.backups WHERE id IN ['fe5dfd90-241a-4afe-864e-d3f2742a4a8b', 'e0325bef-595b-4312-a9a2-bbadb7f61ba0'] AND status == 'BACKUP_CREATED' on node2 http://localhost:None "GET /v1.46/containers/ed245379c226a978f2f324c913afd8945dc62058cf276e232aed9bb131105a78/json HTTP/1.1" 200 None Executing query SELECT name FROM system.parts WHERE database = 'db1' AND table = 'test_table' on node5 run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] http://localhost:None "GET /v1.46/containers/d86a3a10bee16809ac83c2260bef25842cac0f676a7c8b3c2e694f9fbc6490f8/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/ed245379c226a978f2f324c913afd8945dc62058cf276e232aed9bb131105a78/json HTTP/1.1" 200 None run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select dictGet('LocalClickHouse_complex_key_cache_', 'UInt32_', tuple(toUInt64(1),toString('world'))) on local_node Executing query rename table system.query_log to system.query_log_5_test on node2 No clickhouse process running. Start new one. http://localhost:None "POST /v1.46/containers/roottestdatabasebackup-gw1-instance-1/exec HTTP/1.1" 201 74 http://localhost:None "POST /v1.46/exec/45c776abe1b5e23e0b77df61c7b51f1a3f9b2ee6863e601202bc2e10616c3579/start HTTP/1.1" 200 0 http://localhost:None "GET /v1.46/exec/45c776abe1b5e23e0b77df61c7b51f1a3f9b2ee6863e601202bc2e10616c3579/json HTTP/1.1" 200 586 run container_id:roottestallowfeaturetier-gw8-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestallowfeaturetier-gw8-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] http://localhost:None "GET /v1.46/containers/d86a3a10bee16809ac83c2260bef25842cac0f676a7c8b3c2e694f9fbc6490f8/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/ed245379c226a978f2f324c913afd8945dc62058cf276e232aed9bb131105a78/json HTTP/1.1" 200 None ClickHouse node2 started Executing query SELECT hostName() FROM clusterAllReplicas('one_shard_two_nodes', system.one) ORDER BY ALL on node1 Executing query SELECT * FROM test.tbl2 ORDER BY p on node run container_id:roottestallowfeaturetier-gw8-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestallowfeaturetier-gw8-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] No clickhouse process running. Start new one. http://localhost:None "POST /v1.46/containers/roottestallowfeaturetier-gw8-instance-1/exec HTTP/1.1" 201 74 http://localhost:None "GET /v1.46/containers/d86a3a10bee16809ac83c2260bef25842cac0f676a7c8b3c2e694f9fbc6490f8/json HTTP/1.1" 200 None http://localhost:None "POST /v1.46/exec/1cfb4a3e3d198c619f2ae64811b98521c33c259cef0780c49abc614e748ca4c9/start HTTP/1.1" 200 0 http://localhost:None "GET /v1.46/exec/1cfb4a3e3d198c619f2ae64811b98521c33c259cef0780c49abc614e748ca4c9/json HTTP/1.1" 200 586 Executing query SELECT count() FROM system.backups WHERE id IN ['fe5dfd90-241a-4afe-864e-d3f2742a4a8b', 'e0325bef-595b-4312-a9a2-bbadb7f61ba0'] AND status == 'BACKUP_FAILED' on node1 Executing query SELECT name FROM system.parts_columns WHERE database = 'db1' AND table = 'test_table' on node5 http://localhost:None "GET /v1.46/containers/d86a3a10bee16809ac83c2260bef25842cac0f676a7c8b3c2e694f9fbc6490f8/json HTTP/1.1" 200 None Can't connect to MySQL 8 (2003, "Can't connect to MySQL server on '172.16.10.2' ([Errno 111] Connection refused)") Executing query system flush logs on node2 Executing query select dictGetUInt32('LocalClickHouse_complex_key_cache_', 'UInt32_', tuple(toUInt64(1),toString('world'))) on local_node Executing query DROP TABLE IF EXISTS test.tbl2 on node http://localhost:None "GET /v1.46/containers/d86a3a10bee16809ac83c2260bef25842cac0f676a7c8b3c2e694f9fbc6490f8/json HTTP/1.1" 200 None Executing query SELECT hostName(), * FROM clusterAllReplicas("one_shard_two_nodes", system.one) ORDER BY ALL on node1 http://localhost:None "GET /v1.46/containers/d86a3a10bee16809ac83c2260bef25842cac0f676a7c8b3c2e694f9fbc6490f8/json HTTP/1.1" 200 None Executing query SELECT count() FROM system.backups WHERE id IN ['fe5dfd90-241a-4afe-864e-d3f2742a4a8b', 'e0325bef-595b-4312-a9a2-bbadb7f61ba0'] AND status == 'BACKUP_FAILED' on node2 Executing query SELECT name FROM system.databases WHERE name = 'db1' on node5 Executing query select dictHas('LocalClickHouse_complex_key_cache_', tuple(toUInt64(1),toString('world'))) on local_node http://localhost:None "GET /v1.46/containers/d86a3a10bee16809ac83c2260bef25842cac0f676a7c8b3c2e694f9fbc6490f8/json HTTP/1.1" 200 None [gw9] PASSED test_backup_restore/test.py::test_attach_partition Executing query CREATE TABLE test.tbl3 AS test.tbl on node test_backup_restore/test.py::test_replace_partition http://localhost:None "GET /v1.46/containers/d86a3a10bee16809ac83c2260bef25842cac0f676a7c8b3c2e694f9fbc6490f8/json HTTP/1.1" 200 None Executing query SELECT hostName() FROM cluster('one_shard_two_nodes', system.one) on node1 Can't connect to MySQL 8 (2003, "Can't connect to MySQL server on '172.16.10.2' ([Errno 111] Connection refused)") http://localhost:None "GET /v1.46/containers/d86a3a10bee16809ac83c2260bef25842cac0f676a7c8b3c2e694f9fbc6490f8/json HTTP/1.1" 200 None Executing query DROP TABLE tbl ON CLUSTER 'cluster' SYNC on node1 Executing query select dictGetOrDefault('LocalClickHouse_complex_key_cache_', 'UInt32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt32(77)) on local_node Executing query INSERT INTO test.tbl3(p, k) VALUES(toDate(3), 3) on node Executing query SELECT name FROM system.tables WHERE database = 'db1' AND name = 'test_table' on node5 http://localhost:None "GET /v1.46/containers/d86a3a10bee16809ac83c2260bef25842cac0f676a7c8b3c2e694f9fbc6490f8/json HTTP/1.1" 200 None Executing query SELECT hostName() FROM cluster('one_shard_two_nodes', system.one) on node2 http://localhost:None "GET /v1.46/containers/d86a3a10bee16809ac83c2260bef25842cac0f676a7c8b3c2e694f9fbc6490f8/json HTTP/1.1" 200 None run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] http://localhost:None "GET /v1.46/containers/d86a3a10bee16809ac83c2260bef25842cac0f676a7c8b3c2e694f9fbc6490f8/json HTTP/1.1" 200 None Stdout:795 Clickhouse process running. run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select dictGetUInt32OrDefault('LocalClickHouse_complex_key_cache_', 'UInt32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt32(77)) on local_node Executing query INSERT INTO test.tbl3(p, k) VALUES(toDate(4), 4) on node Stdout:795 Executing query SELECT name FROM system.columns WHERE database = 'db1' AND table = 'test_table' on node5 Executing query select 20 on instance run container_id:roottestallowfeaturetier-gw8-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestallowfeaturetier-gw8-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] http://localhost:None "GET /v1.46/containers/d86a3a10bee16809ac83c2260bef25842cac0f676a7c8b3c2e694f9fbc6490f8/json HTTP/1.1" 200 None Executing query select count() from system.tables where database = 'system' and name in ['query_log', 'text_log'] on node2 [gw4] PASSED test_cluster_all_replicas/test.py::test_cluster Executing query SELECT hostName() FROM clusterAllReplicas('one_shard_three_nodes', system.one) ORDER BY ALL settings skip_unavailable_shards=0 on node1 Stdout:775 test_cluster_all_replicas/test.py::test_error_on_unavailable_replica[one_shard_three_nodes] Clickhouse process running. run container_id:roottestallowfeaturetier-gw8-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestallowfeaturetier-gw8-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Can't connect to MySQL 8 (2003, "Can't connect to MySQL server on '172.16.10.2' ([Errno 111] Connection refused)") http://localhost:None "GET /v1.46/containers/d86a3a10bee16809ac83c2260bef25842cac0f676a7c8b3c2e694f9fbc6490f8/json HTTP/1.1" 200 None Stdout:775 Executing query select 20 on instance Executing query select dictGet('LocalClickHouse_complex_key_cache_', 'UInt64_', tuple(toUInt64(1),toString('world'))) on local_node http://localhost:None "GET /v1.46/containers/d86a3a10bee16809ac83c2260bef25842cac0f676a7c8b3c2e694f9fbc6490f8/json HTTP/1.1" 200 None Executing query INSERT INTO test.tbl3(p, k) VALUES(toDate(33), 33) on node Executing query SELECT name FROM system.parts WHERE database = 'db1' AND table = 'test_table' on node5 Executing query RESTORE TABLE tbl FROM Disk('backups', '1') on node1 http://localhost:None "GET /v1.46/containers/d86a3a10bee16809ac83c2260bef25842cac0f676a7c8b3c2e694f9fbc6490f8/json HTTP/1.1" 200 None Executing query rename table system.text_log to system.text_log_6_test on node2 http://localhost:None "GET /v1.46/containers/d86a3a10bee16809ac83c2260bef25842cac0f676a7c8b3c2e694f9fbc6490f8/json HTTP/1.1" 200 None Executing query select dictGetUInt64('LocalClickHouse_complex_key_cache_', 'UInt64_', tuple(toUInt64(1),toString('world'))) on local_node http://localhost:None "GET /v1.46/containers/d86a3a10bee16809ac83c2260bef25842cac0f676a7c8b3c2e694f9fbc6490f8/json HTTP/1.1" 200 None Executing query INSERT INTO test.tbl3(p, k) VALUES(toDate(34), 34) on node Executing query rename table system.query_log to system.query_log_6_test on node2 Executing query SELECT name FROM system.parts_columns WHERE database = 'db1' AND table = 'test_table' on node5 Can't connect to MySQL 8 (2003, "Can't connect to MySQL server on '172.16.10.2' ([Errno 111] Connection refused)") http://localhost:None "GET /v1.46/containers/d86a3a10bee16809ac83c2260bef25842cac0f676a7c8b3c2e694f9fbc6490f8/json HTTP/1.1" 200 None Executing query SELECT * FROM tbl on node1 Executing query select 20 on instance http://localhost:None "GET /v1.46/containers/d86a3a10bee16809ac83c2260bef25842cac0f676a7c8b3c2e694f9fbc6490f8/json HTTP/1.1" 200 None Executing query select dictHas('LocalClickHouse_complex_key_cache_', tuple(toUInt64(1),toString('world'))) on local_node Executing query system flush logs on node2 Executing query select 20 on instance Executing query SELECT * FROM test.tbl3 ORDER BY p on node http://localhost:None "GET /v1.46/containers/d86a3a10bee16809ac83c2260bef25842cac0f676a7c8b3c2e694f9fbc6490f8/json HTTP/1.1" 200 None ClickHouse node started Preparing RedisSimple_flat_UInt8_ Prepared RedisSimple_flat_UInt8_ Preparing RedisSimple_hashed_UInt8_ Prepared RedisSimple_hashed_UInt8_ Preparing RedisSimple_cache_UInt8_ Prepared RedisSimple_cache_UInt8_ Preparing RedisSimple_direct_UInt8_ Prepared RedisSimple_direct_UInt8_ Preparing RedisHash_complex_key_hashed_UInt8_ Prepared RedisHash_complex_key_hashed_UInt8_ Preparing RedisHash_complex_key_cache_UInt8_ Prepared RedisHash_complex_key_cache_UInt8_ Preparing RedisHash_complex_key_direct_UInt8_ Prepared RedisHash_complex_key_direct_UInt8_ Preparing RedisSimple_flat_UInt16_ Prepared RedisSimple_flat_UInt16_ Preparing RedisSimple_hashed_UInt16_ Prepared RedisSimple_hashed_UInt16_ Preparing RedisSimple_cache_UInt16_ Prepared RedisSimple_cache_UInt16_ Preparing RedisSimple_direct_UInt16_ Prepared RedisSimple_direct_UInt16_ Preparing RedisHash_complex_key_hashed_UInt16_ Prepared RedisHash_complex_key_hashed_UInt16_ Preparing RedisHash_complex_key_cache_UInt16_ Prepared RedisHash_complex_key_cache_UInt16_ Preparing RedisHash_complex_key_direct_UInt16_ Prepared RedisHash_complex_key_direct_UInt16_ Preparing RedisSimple_flat_UInt32_ Prepared RedisSimple_flat_UInt32_ Preparing RedisSimple_hashed_UInt32_ Prepared RedisSimple_hashed_UInt32_ Preparing RedisSimple_cache_UInt32_ Prepared RedisSimple_cache_UInt32_ Preparing RedisSimple_direct_UInt32_ Prepared RedisSimple_direct_UInt32_ Preparing RedisHash_complex_key_hashed_UInt32_ Prepared RedisHash_complex_key_hashed_UInt32_ Preparing RedisHash_complex_key_cache_UInt32_ Prepared RedisHash_complex_key_cache_UInt32_ Preparing RedisHash_complex_key_direct_UInt32_ Prepared RedisHash_complex_key_direct_UInt32_ Preparing RedisSimple_flat_UInt64_ Prepared RedisSimple_flat_UInt64_ Preparing RedisSimple_hashed_UInt64_ Prepared RedisSimple_hashed_UInt64_ Preparing RedisSimple_cache_UInt64_ Prepared RedisSimple_cache_UInt64_ Preparing RedisSimple_direct_UInt64_ Prepared RedisSimple_direct_UInt64_ Preparing RedisHash_complex_key_hashed_UInt64_ Prepared RedisHash_complex_key_hashed_UInt64_ Preparing RedisHash_complex_key_cache_UInt64_ Prepared RedisHash_complex_key_cache_UInt64_ Preparing RedisHash_complex_key_direct_UInt64_ Prepared RedisHash_complex_key_direct_UInt64_ Preparing RedisSimple_flat_Int8_ Prepared RedisSimple_flat_Int8_ Preparing RedisSimple_hashed_Int8_ Prepared RedisSimple_hashed_Int8_ Preparing RedisSimple_cache_Int8_ Prepared RedisSimple_cache_Int8_ Preparing RedisSimple_direct_Int8_ Prepared RedisSimple_direct_Int8_ Preparing RedisHash_complex_key_hashed_Int8_ Prepared RedisHash_complex_key_hashed_Int8_ Preparing RedisHash_complex_key_cache_Int8_ Prepared RedisHash_complex_key_cache_Int8_ Preparing RedisHash_complex_key_direct_Int8_ Prepared RedisHash_complex_key_direct_Int8_ Preparing RedisSimple_flat_Int16_ Prepared RedisSimple_flat_Int16_ Preparing RedisSimple_hashed_Int16_ Prepared RedisSimple_hashed_Int16_ Preparing RedisSimple_cache_Int16_ Prepared RedisSimple_cache_Int16_ Preparing RedisSimple_direct_Int16_ Prepared RedisSimple_direct_Int16_ Preparing RedisHash_complex_key_hashed_Int16_ Prepared RedisHash_complex_key_hashed_Int16_ Preparing RedisHash_complex_key_cache_Int16_ Prepared RedisHash_complex_key_cache_Int16_ Preparing RedisHash_complex_key_direct_Int16_ Prepared RedisHash_complex_key_direct_Int16_ Preparing RedisSimple_flat_Int32_ Prepared RedisSimple_flat_Int32_ Preparing RedisSimple_hashed_Int32_ Prepared RedisSimple_hashed_Int32_ Preparing RedisSimple_cache_Int32_ Prepared RedisSimple_cache_Int32_ Preparing RedisSimple_direct_Int32_ Prepared RedisSimple_direct_Int32_ Preparing RedisHash_complex_key_hashed_Int32_ Prepared RedisHash_complex_key_hashed_Int32_ Preparing RedisHash_complex_key_cache_Int32_ Prepared RedisHash_complex_key_cache_Int32_ Preparing RedisHash_complex_key_direct_Int32_ Prepared RedisHash_complex_key_direct_Int32_ Preparing RedisSimple_flat_Int64_ Prepared RedisSimple_flat_Int64_ Preparing RedisSimple_hashed_Int64_ Prepared RedisSimple_hashed_Int64_ Preparing RedisSimple_cache_Int64_ Prepared RedisSimple_cache_Int64_ Preparing RedisSimple_direct_Int64_ Prepared RedisSimple_direct_Int64_ Preparing RedisHash_complex_key_hashed_Int64_ Prepared RedisHash_complex_key_hashed_Int64_ Preparing RedisHash_complex_key_cache_Int64_ Prepared RedisHash_complex_key_cache_Int64_ Preparing RedisHash_complex_key_direct_Int64_ Prepared RedisHash_complex_key_direct_Int64_ Preparing RedisSimple_flat_UUID_ Prepared RedisSimple_flat_UUID_ Preparing RedisSimple_hashed_UUID_ Prepared RedisSimple_hashed_UUID_ Preparing RedisSimple_cache_UUID_ Prepared RedisSimple_cache_UUID_ Preparing RedisSimple_direct_UUID_ Prepared RedisSimple_direct_UUID_ Preparing RedisHash_complex_key_hashed_UUID_ Prepared RedisHash_complex_key_hashed_UUID_ Preparing RedisHash_complex_key_cache_UUID_ Prepared RedisHash_complex_key_cache_UUID_ Preparing RedisHash_complex_key_direct_UUID_ Prepared RedisHash_complex_key_direct_UUID_ Preparing RedisSimple_flat_Date_ Prepared RedisSimple_flat_Date_ Preparing RedisSimple_hashed_Date_ Prepared RedisSimple_hashed_Date_ Preparing RedisSimple_cache_Date_ Prepared RedisSimple_cache_Date_ Preparing RedisSimple_direct_Date_ Prepared RedisSimple_direct_Date_ Preparing RedisHash_complex_key_hashed_Date_ Prepared RedisHash_complex_key_hashed_Date_ Preparing RedisHash_complex_key_cache_Date_ Prepared RedisHash_complex_key_cache_Date_ Preparing RedisHash_complex_key_direct_Date_ Prepared RedisHash_complex_key_direct_Date_ Preparing RedisSimple_flat_DateTime_ Prepared RedisSimple_flat_DateTime_ Preparing RedisSimple_hashed_DateTime_ Prepared RedisSimple_hashed_DateTime_ Preparing RedisSimple_cache_DateTime_ Prepared RedisSimple_cache_DateTime_ Preparing RedisSimple_direct_DateTime_ Prepared RedisSimple_direct_DateTime_ Preparing RedisHash_complex_key_hashed_DateTime_ Prepared RedisHash_complex_key_hashed_DateTime_ Preparing RedisHash_complex_key_cache_DateTime_ Prepared RedisHash_complex_key_cache_DateTime_ Preparing RedisHash_complex_key_direct_DateTime_ Prepared RedisHash_complex_key_direct_DateTime_ Preparing RedisSimple_flat_String_ Prepared RedisSimple_flat_String_ Preparing RedisSimple_hashed_String_ Prepared RedisSimple_hashed_String_ Preparing RedisSimple_cache_String_ Prepared RedisSimple_cache_String_ Preparing RedisSimple_direct_String_ Prepared RedisSimple_direct_String_ Preparing RedisHash_complex_key_hashed_String_ Prepared RedisHash_complex_key_hashed_String_ [gw6] PASSED test_config_substitutions/test.py::test_allow_databases Preparing RedisHash_complex_key_cache_String_ Prepared RedisHash_complex_key_cache_String_ Executing query select value from system.settings where name = 'max_query_size' on node1 Preparing RedisHash_complex_key_direct_String_ Prepared RedisHash_complex_key_direct_String_ Preparing RedisSimple_flat_Float32_ Prepared RedisSimple_flat_Float32_ Preparing RedisSimple_hashed_Float32_ Prepared RedisSimple_hashed_Float32_ Preparing RedisSimple_cache_Float32_ Prepared RedisSimple_cache_Float32_ Preparing RedisSimple_direct_Float32_ Prepared RedisSimple_direct_Float32_ Preparing RedisHash_complex_key_hashed_Float32_ Prepared RedisHash_complex_key_hashed_Float32_ Preparing RedisHash_complex_key_cache_Float32_ Prepared RedisHash_complex_key_cache_Float32_ Preparing RedisHash_complex_key_direct_Float32_ test_config_substitutions/test.py::test_config Prepared RedisHash_complex_key_direct_Float32_ Preparing RedisSimple_flat_Float64_ Prepared RedisSimple_flat_Float64_ Preparing RedisSimple_hashed_Float64_ Prepared RedisSimple_hashed_Float64_ Preparing RedisSimple_cache_Float64_ Prepared RedisSimple_cache_Float64_ Preparing RedisSimple_direct_Float64_ Prepared RedisSimple_direct_Float64_ Preparing RedisHash_complex_key_hashed_Float64_ Prepared RedisHash_complex_key_hashed_Float64_ Preparing RedisHash_complex_key_cache_Float64_ Prepared RedisHash_complex_key_cache_Float64_ Preparing RedisHash_complex_key_direct_Float64_ Prepared RedisHash_complex_key_direct_Float64_ Run test with id: 10 Executing query system reload dictionaries on node Executing query DROP TABLE IF EXISTS tbl ON CLUSTER 'cluster3' SYNC on node1 [gw2] PASSED test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[http-False] Executing query select dictGetOrDefault('LocalClickHouse_complex_key_cache_', 'UInt64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt64(88)) on local_node Can't connect to MySQL 8 (2003, "Can't connect to MySQL server on '172.16.10.2' ([Errno 111] Connection refused)") copy from /var/lib/clickhouse/shadow/*/data/test/tbl to /var/lib/clickhouse/data/test/tbl3/detached run container_id:roottestbackuprestore-gw9-node-1 detach:False nothrow:False cmd: ['bash', '-c', 'cp -r /var/lib/clickhouse/shadow/*/data/test/tbl -T /var/lib/clickhouse/data/test/tbl3/detached'] Command:[docker exec roottestbackuprestore-gw9-node-1 bash -c cp -r /var/lib/clickhouse/shadow/*/data/test/tbl -T /var/lib/clickhouse/data/test/tbl3/detached] Executing query ALTER TABLE test.tbl3 REPLACE PARTITION 197002 FROM test.tbl on node Executing query select value from system.settings where name = 'max_query_size' on node2 Executing query select 20 on instance Executing query select dictGetUInt64OrDefault('LocalClickHouse_complex_key_cache_', 'UInt64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt64(88)) on local_node Executing query SELECT sleep(2) on node Executing query DROP TABLE IF EXISTS tbl2 ON CLUSTER 'cluster3' SYNC on node1 Executing query select 20 on instance Can't connect to MySQL 8 (2003, "Can't connect to MySQL server on '172.16.10.2' ([Errno 111] Connection refused)") Executing query select value from system.settings where name = 'max_query_size' on node3 Executing query select dictGet('LocalClickHouse_complex_key_cache_', 'Int8_', tuple(toUInt64(1),toString('world'))) on local_node Executing query select count() from system.tables where database = 'system' and name in ['query_log', 'text_log'] on node2 Executing query DROP DATABASE IF EXISTS mydb ON CLUSTER 'cluster3' SYNC on node1 Executing query select value from system.settings where name = 'max_query_size' on node4 Executing query select 20 on instance Executing query select dictGetInt8('LocalClickHouse_complex_key_cache_', 'Int8_', tuple(toUInt64(1),toString('world'))) on local_node Executing query rename table system.text_log to system.text_log_7_test on node2 Can't connect to MySQL 8 (2003, "Can't connect to MySQL server on '172.16.10.2' ([Errno 111] Connection refused)") Executing query select 20 on instance Executing query select value from system.settings where name = 'max_query_size' on node6 Executing query DROP DATABASE IF EXISTS mydb2 ON CLUSTER 'cluster3' SYNC on node1 Executing query rename table system.query_log to system.query_log_7_test on node2 Executing query SELECT id, value FROM test_database_backup.test_table_1 on instance Executing query select dictHas('LocalClickHouse_complex_key_cache_', tuple(toUInt64(1),toString('world'))) on local_node Executing query system reload dictionary RedisSimple_flat_DateTime_ on node Executing query SELECT engine_full FROM system.tables WHERE name = 'test_experimental' on instance Executing query system flush logs on node2 Executing query SELECT id, value FROM test_database_backup.test_table_2 on instance Can't connect to MySQL 8 (2003, "Can't connect to MySQL server on '172.16.10.2' ([Errno 111] Connection refused)") Executing query select value from system.settings where name = 'max_query_size' on node7 Executing query DROP USER IF EXISTS u1, u2 ON CLUSTER 'cluster3' on node1 Executing query select dictGet('RedisSimple_flat_DateTime_', 'DateTime_', toUInt64(1)) on node Executing query select dictGetOrDefault('LocalClickHouse_complex_key_cache_', 'Int8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt8(-55)) on local_node Executing query CREATE TABLE test_experimental_new (uid String, version UInt32, is_deleted UInt8) ENGINE = ReplacingMergeTree(version, is_deleted) ORDER by (uid) SETTINGS allow_experimental_replacing_merge_with_cleanup=1; on instance Executing query SELECT id, value FROM test_database_backup.test_table_3 on instance Executing query select value from system.settings where name = 'max_threads' on node7 Executing query select dictGetDateTime('RedisSimple_flat_DateTime_', 'DateTime_', toUInt64(1)) on node Executing query select dictGetInt8OrDefault('LocalClickHouse_complex_key_cache_', 'Int8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt8(-55)) on local_node Executing query CREATE TABLE tbl ON CLUSTER 'cluster' (x UInt8) ENGINE=ReplicatedMergeTree('/clickhouse/tables/tbl/', '{replica}')ORDER BY x on node1 test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[http-True] Can't connect to MySQL 8 (2003, "Can't connect to MySQL server on '172.16.10.2' ([Errno 111] Connection refused)") Executing query CREATE TABLE test_experimental_new (uid String, version UInt32, is_deleted UInt8) ENGINE = ReplacingMergeTree(version, is_deleted) ORDER by (uid); on instance Executing query DROP DATABASE test_database_backup on instance Executing query select value from system.settings where name = 'max_query_size' on node8 Executing query select dictHas('RedisSimple_flat_DateTime_', toUInt64(1)) on node Executing query select dictGet('LocalClickHouse_complex_key_cache_', 'Int16_', tuple(toUInt64(1),toString('world'))) on local_node Executing query ALTER TABLE test_experimental_new MODIFY setting allow_experimental_replacing_merge_with_cleanup=1 on instance Executing query DROP DATABASE test_database on instance Executing query INSERT INTO tbl VALUES (1) on node1 Executing query select dictGetOrDefault('RedisSimple_flat_DateTime_', 'DateTime_', toUInt64(9999999), toDateTime('2018-12-30 00:00:00')) on node Executing query select dictGetInt16('LocalClickHouse_complex_key_cache_', 'Int16_', tuple(toUInt64(1),toString('world'))) on local_node [gw6] PASSED test_config_substitutions/test.py::test_config test_config_substitutions/test.py::test_config_from_env_overrides run container_id:roottestconfigsubstitutions-gw6-node7-1 detach:False nothrow:False cmd: ['bash', '-c', 'test ! -f /etc/clickhouse-server/users.d/000-users_with_env_subst.xml || mv --no-clobber /etc/clickhouse-server/users.d/000-users_with_env_subst.xml /tmp/000-users_with_env_subst_6ce0fd54-ad41-4441-a9c7-f99d0b98c65b.xml'] Command:[docker exec roottestconfigsubstitutions-gw6-node7-1 bash -c test ! -f /etc/clickhouse-server/users.d/000-users_with_env_subst.xml || mv --no-clobber /etc/clickhouse-server/users.d/000-users_with_env_subst.xml /tmp/000-users_with_env_subst_6ce0fd54-ad41-4441-a9c7-f99d0b98c65b.xml] run container_id:roottestconfigsubstitutions-gw6-node7-1 detach:False nothrow:False cmd: ['bash', '-c', 'echo \'\n\n \n \n \n 100\n \n \n \n \n \n default\n default\n \n\n \n \n \n\n\' > /etc/clickhouse-server/users.d/000-users_with_env_subst.xml'] Command:[docker exec roottestconfigsubstitutions-gw6-node7-1 bash -c echo ' 100 default default ' > /etc/clickhouse-server/users.d/000-users_with_env_subst.xml] Executing query select count() from system.tables where database = 'system' and name in ['query_log', 'text_log'] on node2 Can't connect to MySQL 8 (2003, "Can't connect to MySQL server on '172.16.10.2' ([Errno 111] Connection refused)") Executing query SYSTEM RELOAD CONFIG on node7 Executing query SELECT * FROM test.tbl3 ORDER BY p on node run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', 'rm -rf /backups/'] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c rm -rf /backups/] Executing query BACKUP TABLE tbl ON CLUSTER 'cluster' TO Disk('backups', '2') ASYNC on node1 via HTTP interface Starting new HTTP connection (1): 172.16.5.5:8123 Executing query DROP TABLE IF EXISTS test_experimental_new on instance Executing query select dictGetDateTimeOrDefault('RedisSimple_flat_DateTime_', 'DateTime_', toUInt64(9999999), toDateTime('2018-12-30 00:00:00')) on node http://172.16.5.5:8123 "GET /?query=BACKUP+TABLE+tbl+ON+CLUSTER+%27cluster%27+TO+Disk%28%27backups%27%2C+%272%27%29+ASYNC HTTP/1.1" 200 None Executing query BACKUP TABLE tbl ON CLUSTER 'cluster' TO Disk('backups', '2') ASYNC on node2 via HTTP interface Executing query select dictHas('LocalClickHouse_complex_key_cache_', tuple(toUInt64(1),toString('world'))) on local_node Starting new HTTP connection (1): 172.16.5.6:8123 http://172.16.5.6:8123 "GET /?query=BACKUP+TABLE+tbl+ON+CLUSTER+%27cluster%27+TO+Disk%28%27backups%27%2C+%272%27%29+ASYNC HTTP/1.1" 200 None Executing query SELECT status FROM system.backups WHERE id IN ['05b14057-7353-4de3-befb-b3432c6be3d3', '32721448-aed9-4566-9809-6c9d91131896'] AND status == 'CREATING_BACKUP' on node1 run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', 'rm -rf /local_plain/'] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c rm -rf /local_plain/] http://172.16.8.4:9001 "GET /root?delimiter=&encoding-type=url&list-type=2&max-keys=1000&prefix= HTTP/1.1" 200 0 run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', 'rm -rf /backups/'] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c rm -rf /backups/] [gw1] PASSED test_database_backup/test.py::test_database_backup_database[Disk('backup_disk_local', 'test_database_backup')] test_database_backup/test.py::test_database_backup_database[Disk('backup_disk_object_storage_local_plain', 'test_database_backup')] Executing query rename table system.text_log to system.text_log_8_test on node2 run container_id:roottestallowfeaturetier-gw8-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "sed -i 's/1/0/g' /etc/clickhouse-server/config.d/allow_feature_tier.xml"] Command:[docker exec roottestallowfeaturetier-gw8-instance-1 bash -c sed -i 's/1/0/g' /etc/clickhouse-server/config.d/allow_feature_tier.xml] Can't connect to MySQL 8 (2003, "Can't connect to MySQL server on '172.16.10.2' ([Errno 111] Connection refused)") Executing query ALTER TABLE test.tbl3 UPDATE k=10 WHERE 1 on node run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', 'rm -rf /local_plain/'] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c rm -rf /local_plain/] Executing query select dictGet('RedisSimple_flat_DateTime_', 'DateTime_', toUInt64(2)) on node Executing query select dictGetOrDefault('LocalClickHouse_complex_key_cache_', 'Int16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt16(-66)) on local_node Executing query SYSTEM RELOAD CONFIG on instance run container_id:roottestconfigsubstitutions-gw6-node7-1 detach:False nothrow:False cmd: ['bash', '-c', 'test ! -f /tmp/000-users_with_env_subst_6ce0fd54-ad41-4441-a9c7-f99d0b98c65b.xml || mv /tmp/000-users_with_env_subst_6ce0fd54-ad41-4441-a9c7-f99d0b98c65b.xml /etc/clickhouse-server/users.d/000-users_with_env_subst.xml'] Command:[docker exec roottestconfigsubstitutions-gw6-node7-1 bash -c test ! -f /tmp/000-users_with_env_subst_6ce0fd54-ad41-4441-a9c7-f99d0b98c65b.xml || mv /tmp/000-users_with_env_subst_6ce0fd54-ad41-4441-a9c7-f99d0b98c65b.xml /etc/clickhouse-server/users.d/000-users_with_env_subst.xml] http://172.16.8.4:9001 "GET /root?delimiter=&encoding-type=url&list-type=2&max-keys=1000&prefix= HTTP/1.1" 200 0 Executing query DROP DATABASE IF EXISTS test_database; DROP DATABASE IF EXISTS test_database_backup; CREATE DATABASE test_database; CREATE TABLE test_database.test_table_1 (id UInt64, value String) ENGINE=MergeTree ORDER BY id; INSERT INTO test_database.test_table_1 VALUES (0, 'test_database.test_table_1'); CREATE TABLE test_database.test_table_2 (id UInt64, value String) ENGINE=MergeTree ORDER BY id; INSERT INTO test_database.test_table_2 VALUES (0, 'test_database.test_table_2'); CREATE TABLE test_database.test_table_3 (id UInt64, value String) ENGINE=MergeTree ORDER BY id; INSERT INTO test_database.test_table_3 VALUES (0, 'test_database.test_table_3'); BACKUP DATABASE test_database TO Disk('backup_disk_object_storage_local_plain', 'test_database_backup'); CREATE DATABASE test_database_backup ENGINE = Backup('test_database', Disk('backup_disk_object_storage_local_plain', 'test_database_backup')); on instance Executing query rename table system.query_log to system.query_log_8_test on node2 run container_id:roottestconfigsubstitutions-gw6-node7-1 detach:False nothrow:False cmd: ['bash', '-c', 'test ! -f /etc/clickhouse-server/users.d/000-users_with_env_subst.xml || mv --no-clobber /etc/clickhouse-server/users.d/000-users_with_env_subst.xml /tmp/000-users_with_env_subst_161e6c53-b101-4241-98e4-5f3598cd21f4.xml'] Command:[docker exec roottestconfigsubstitutions-gw6-node7-1 bash -c test ! -f /etc/clickhouse-server/users.d/000-users_with_env_subst.xml || mv --no-clobber /etc/clickhouse-server/users.d/000-users_with_env_subst.xml /tmp/000-users_with_env_subst_161e6c53-b101-4241-98e4-5f3598cd21f4.xml] Executing query SELECT sleep(2) on node run container_id:roottestconfigsubstitutions-gw6-node7-1 detach:False nothrow:False cmd: ['bash', '-c', 'echo \'\n\n \n \n \n 1\n \n \n \n \n \n default\n default\n \n\n \n \n \n\n\' > /etc/clickhouse-server/users.d/000-users_with_env_subst.xml'] Command:[docker exec roottestconfigsubstitutions-gw6-node7-1 bash -c echo ' 1 default default ' > /etc/clickhouse-server/users.d/000-users_with_env_subst.xml] Executing query select dictGetInt16OrDefault('LocalClickHouse_complex_key_cache_', 'Int16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt16(-66)) on local_node Executing query select dictGetDateTime('RedisSimple_flat_DateTime_', 'DateTime_', toUInt64(2)) on node Can't connect to MySQL 8 (2003, "Can't connect to MySQL server on '172.16.10.2' ([Errno 111] Connection refused)") Executing query system flush logs on node2 Executing query SELECT value FROM system.server_settings where name = 'allow_feature_tier' on instance Executing query SYSTEM RELOAD CONFIG on node7 Executing query select dictGet('LocalClickHouse_complex_key_cache_', 'Int32_', tuple(toUInt64(1),toString('world'))) on local_node Executing query select dictHas('RedisSimple_flat_DateTime_', toUInt64(2)) on node Executing query SELECT status FROM system.backups WHERE id IN ['05b14057-7353-4de3-befb-b3432c6be3d3', '32721448-aed9-4566-9809-6c9d91131896'] AND status == 'CREATING_BACKUP' on node1 Executing query DROP TABLE IF EXISTS test_experimental on instance Executing query SELECT id, value FROM test_database_backup.test_table_1 on instance Can't connect to MySQL 8 (2003, "Can't connect to MySQL server on '172.16.10.2' ([Errno 111] Connection refused)") Executing query select dictGetOrDefault('RedisSimple_flat_DateTime_', 'DateTime_', toUInt64(9999999), toDateTime('2018-12-30 00:00:00')) on node Executing query select dictGetInt32('LocalClickHouse_complex_key_cache_', 'Int32_', tuple(toUInt64(1),toString('world'))) on local_node run container_id:roottestconfigsubstitutions-gw6-node7-1 detach:False nothrow:False cmd: ['bash', '-c', 'test ! -f /tmp/000-users_with_env_subst_161e6c53-b101-4241-98e4-5f3598cd21f4.xml || mv /tmp/000-users_with_env_subst_161e6c53-b101-4241-98e4-5f3598cd21f4.xml /etc/clickhouse-server/users.d/000-users_with_env_subst.xml'] Command:[docker exec roottestconfigsubstitutions-gw6-node7-1 bash -c test ! -f /tmp/000-users_with_env_subst_161e6c53-b101-4241-98e4-5f3598cd21f4.xml || mv /tmp/000-users_with_env_subst_161e6c53-b101-4241-98e4-5f3598cd21f4.xml /etc/clickhouse-server/users.d/000-users_with_env_subst.xml] [gw8] PASSED test_allow_feature_tier/test.py::test_allow_feature_tier_in_mergetree_settings Executing query SELECT value FROM system.server_settings where name = 'allow_feature_tier' on instance test_allow_feature_tier/test.py::test_allow_feature_tier_in_mergetree_settings_with_old_compatibility Executing query SYSTEM RELOAD CONFIG on node7 [gw6] PASSED test_config_substitutions/test.py::test_config_from_env_overrides test_config_substitutions/test.py::test_config_merge_from_env_overrides Executing query SELECT id, value FROM test_database_backup.test_table_2 on instance Executing query DROP TABLE IF EXISTS test_experimental on instance Executing query select dictGetDateTimeOrDefault('RedisSimple_flat_DateTime_', 'DateTime_', toUInt64(9999999), toDateTime('2018-12-30 00:00:00')) on node Executing query select dictHas('LocalClickHouse_complex_key_cache_', tuple(toUInt64(1),toString('world'))) on local_node Executing query SELECT value FROM system.server_settings WHERE name='max_thread_pool_size' on node7 Can't connect to MySQL 8 (2003, "Can't connect to MySQL server on '172.16.10.2' ([Errno 111] Connection refused)") Executing query SELECT id, value FROM test_database_backup.test_table_3 on instance Executing query select count() from system.tables where database = 'system' and name in ['query_log', 'text_log'] on node2 Executing query SELECT status FROM system.backups WHERE id IN ['05b14057-7353-4de3-befb-b3432c6be3d3', '32721448-aed9-4566-9809-6c9d91131896'] AND status == 'CREATING_BACKUP' on node1 run container_id:roottestallowfeaturetier-gw8-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "sed -i 's/0/1/g' /etc/clickhouse-server/config.d/allow_feature_tier.xml"] Command:[docker exec roottestallowfeaturetier-gw8-instance-1 bash -c sed -i 's/0/1/g' /etc/clickhouse-server/config.d/allow_feature_tier.xml] Executing query SYSTEM RELOAD CONFIG on instance Executing query select dictGetOrDefault('LocalClickHouse_complex_key_cache_', 'Int32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt32(-77)) on local_node Executing query system reload dictionary RedisSimple_hashed_DateTime_ on node run container_id:roottestconfigsubstitutions-gw6-node7-1 detach:False nothrow:False cmd: ['bash', '-c', 'test ! -f /etc/clickhouse-server/config.d/010-server_with_env_subst.xml || mv --no-clobber /etc/clickhouse-server/config.d/010-server_with_env_subst.xml /tmp/010-server_with_env_subst_4ca7d27f-f1c0-40a7-97ad-1f38e1cb412f.xml'] Command:[docker exec roottestconfigsubstitutions-gw6-node7-1 bash -c test ! -f /etc/clickhouse-server/config.d/010-server_with_env_subst.xml || mv --no-clobber /etc/clickhouse-server/config.d/010-server_with_env_subst.xml /tmp/010-server_with_env_subst_4ca7d27f-f1c0-40a7-97ad-1f38e1cb412f.xml] run container_id:roottestconfigsubstitutions-gw6-node7-1 detach:False nothrow:False cmd: ['bash', '-c', 'echo \'\n\n 9000\n\n\' > /etc/clickhouse-server/config.d/010-server_with_env_subst.xml'] Command:[docker exec roottestconfigsubstitutions-gw6-node7-1 bash -c echo ' 9000 ' > /etc/clickhouse-server/config.d/010-server_with_env_subst.xml] run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:True cmd: ['bash', '-c', 'ps -C clickhouse'] Command:[docker exec -u root roottestdatabasebackup-gw1-instance-1 bash -c ps -C clickhouse] Executing query SYSTEM RELOAD CONFIG on node7 Stdout: PID TTY TIME CMD Stdout: 795 ? 00:00:03 clickhouse run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', 'pkill clickhouse'] Command:[docker exec -u root roottestdatabasebackup-gw1-instance-1 bash -c pkill clickhouse] Executing query select dictGet('RedisSimple_hashed_DateTime_', 'DateTime_', toUInt64(1)) on node Executing query SELECT status FROM system.backups WHERE id IN ['05b14057-7353-4de3-befb-b3432c6be3d3', '32721448-aed9-4566-9809-6c9d91131896'] AND status == 'CREATING_BACKUP' on node2 run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select dictGetInt32OrDefault('LocalClickHouse_complex_key_cache_', 'Int32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt32(-77)) on local_node Executing query rename table system.text_log to system.text_log_9_test on node2 Stdout:795 Can't connect to MySQL 8 (2003, "Can't connect to MySQL server on '172.16.10.2' ([Errno 111] Connection refused)") Executing query SELECT value FROM system.server_settings where name = 'allow_feature_tier' on instance run container_id:roottestconfigsubstitutions-gw6-node7-1 detach:False nothrow:False cmd: ['bash', '-c', 'test ! -f /tmp/010-server_with_env_subst_4ca7d27f-f1c0-40a7-97ad-1f38e1cb412f.xml || mv /tmp/010-server_with_env_subst_4ca7d27f-f1c0-40a7-97ad-1f38e1cb412f.xml /etc/clickhouse-server/config.d/010-server_with_env_subst.xml'] Command:[docker exec roottestconfigsubstitutions-gw6-node7-1 bash -c test ! -f /tmp/010-server_with_env_subst_4ca7d27f-f1c0-40a7-97ad-1f38e1cb412f.xml || mv /tmp/010-server_with_env_subst_4ca7d27f-f1c0-40a7-97ad-1f38e1cb412f.xml /etc/clickhouse-server/config.d/010-server_with_env_subst.xml] Executing query select dictGetDateTime('RedisSimple_hashed_DateTime_', 'DateTime_', toUInt64(1)) on node run container_id:roottestconfigsubstitutions-gw6-node3-1 detach:False nothrow:True cmd: ['bash', '-c', 'ps -C clickhouse'] Command:[docker exec -u root roottestconfigsubstitutions-gw6-node3-1 bash -c ps -C clickhouse] [gw6] PASSED test_config_substitutions/test.py::test_config_merge_from_env_overrides test_config_substitutions/test.py::test_config_multiple_zk_substitutions Executing query rename table system.query_log to system.query_log_9_test on node2 Executing query select dictGet('LocalClickHouse_complex_key_cache_', 'Int64_', tuple(toUInt64(1),toString('world'))) on local_node Stdout: PID TTY TIME CMD Stdout: 8 ? 00:00:04 clickhouse run container_id:roottestconfigsubstitutions-gw6-node3-1 detach:False nothrow:False cmd: ['bash', '-c', 'pkill clickhouse'] Command:[docker exec -u root roottestconfigsubstitutions-gw6-node3-1 bash -c pkill clickhouse] Executing query SELECT count() FROM system.backups WHERE id IN ['05b14057-7353-4de3-befb-b3432c6be3d3', '32721448-aed9-4566-9809-6c9d91131896'] AND status == 'BACKUP_CREATED' on node1 run container_id:roottestconfigsubstitutions-gw6-node3-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestconfigsubstitutions-gw6-node3-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query create table b (a Int64) ENGINE=MergeTree() order by a; on instance Stdout:8 Executing query system flush logs on node2 [gw4] PASSED test_cluster_all_replicas/test.py::test_error_on_unavailable_replica[one_shard_three_nodes] Executing query SELECT hostName() FROM clusterAllReplicas('two_shards_three_nodes', system.one) ORDER BY ALL settings skip_unavailable_shards=0 on node1 test_cluster_all_replicas/test.py::test_error_on_unavailable_replica[two_shards_three_nodes] Mysql 8 Started ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_dictionaries_all_layouts_separate_sources/_instances-mysql-0-gw3/.env --project-name roottestdictionariesalllayoutsseparatesourcesmysql-gw3 --file /ClickHouse/tests/integration/test_dictionaries_all_layouts_separate_sources/_instances-mysql-0-gw3/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_mysql_8_0.yml up -d --no-recreate') Command:[docker compose --env-file /ClickHouse/tests/integration/test_dictionaries_all_layouts_separate_sources/_instances-mysql-0-gw3/.env --project-name roottestdictionariesalllayoutsseparatesourcesmysql-gw3 --file /ClickHouse/tests/integration/test_dictionaries_all_layouts_separate_sources/_instances-mysql-0-gw3/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_mysql_8_0.yml up -d --no-recreate] Executing query select dictHas('RedisSimple_hashed_DateTime_', toUInt64(1)) on node Executing query select dictGetInt64('LocalClickHouse_complex_key_cache_', 'Int64_', tuple(toUInt64(1),toString('world'))) on local_node run container_id:roottestallowfeaturetier-gw8-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "sed -i 's/1/0/g' /etc/clickhouse-server/config.d/allow_feature_tier.xml"] Command:[docker exec roottestallowfeaturetier-gw8-instance-1 bash -c sed -i 's/1/0/g' /etc/clickhouse-server/config.d/allow_feature_tier.xml] Executing query SELECT count() FROM system.backups WHERE id IN ['05b14057-7353-4de3-befb-b3432c6be3d3', '32721448-aed9-4566-9809-6c9d91131896'] AND status == 'BACKUP_CREATED' on node2 Executing query SELECT * FROM test.tbl3 ORDER BY p on node Executing query SYSTEM RELOAD CONFIG on instance Executing query select dictGetOrDefault('RedisSimple_hashed_DateTime_', 'DateTime_', toUInt64(9999999), toDateTime('2018-12-30 00:00:00')) on node Stderr: Container roottestdictionariesalllayoutsseparatesourcesmysql-gw3-mysql80-1 Running Stderr: Container roottestdictionariesalllayoutsseparatesourcesmysql-gw3-node-1 Creating Stderr: Container roottestdictionariesalllayoutsseparatesourcesmysql-gw3-node-1 Created Stderr: Container roottestdictionariesalllayoutsseparatesourcesmysql-gw3-node-1 Starting Stderr: Container roottestdictionariesalllayoutsseparatesourcesmysql-gw3-node-1 Started ClickHouse instance created get_instance_ip instance_name=node http://localhost:None "GET /v1.46/containers/roottestdictionariesalllayoutsseparatesourcesmysql-gw3-node-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node http://localhost:None "GET /v1.46/containers/roottestdictionariesalllayoutsseparatesourcesmysql-gw3-node-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node, ip: 172.16.10.3... http://localhost:None "GET /v1.46/containers/roottestdictionariesalllayoutsseparatesourcesmysql-gw3-node-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/e5a667165acdebe03a7248a291501be53e66923bfb42e2902d4dda904ff2e4e7/json HTTP/1.1" 200 None Executing query select dictHas('LocalClickHouse_complex_key_cache_', tuple(toUInt64(1),toString('world'))) on local_node run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] http://localhost:None "GET /v1.46/containers/e5a667165acdebe03a7248a291501be53e66923bfb42e2902d4dda904ff2e4e7/json HTTP/1.1" 200 None Stdout:795 Executing query DROP TABLE IF EXISTS test.tbl3 on node Executing query SELECT count() FROM system.backups WHERE id IN ['05b14057-7353-4de3-befb-b3432c6be3d3', '32721448-aed9-4566-9809-6c9d91131896'] AND status == 'BACKUP_FAILED' on node1 Executing query select dictGetDateTimeOrDefault('RedisSimple_hashed_DateTime_', 'DateTime_', toUInt64(9999999), toDateTime('2018-12-30 00:00:00')) on node Executing query SELECT value FROM system.server_settings where name = 'allow_feature_tier' on instance http://localhost:None "GET /v1.46/containers/e5a667165acdebe03a7248a291501be53e66923bfb42e2902d4dda904ff2e4e7/json HTTP/1.1" 200 None Executing query select dictGetOrDefault('LocalClickHouse_complex_key_cache_', 'Int64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt64(-88)) on local_node http://localhost:None "GET /v1.46/containers/e5a667165acdebe03a7248a291501be53e66923bfb42e2902d4dda904ff2e4e7/json HTTP/1.1" 200 None [gw9] PASSED test_backup_restore/test.py::test_replace_partition test_backup_restore/test.py::test_restore Executing query CREATE TABLE test.tbl1 AS test.tbl on node http://localhost:None "GET /v1.46/containers/e5a667165acdebe03a7248a291501be53e66923bfb42e2902d4dda904ff2e4e7/json HTTP/1.1" 200 None Executing query SELECT count() FROM system.backups WHERE id IN ['05b14057-7353-4de3-befb-b3432c6be3d3', '32721448-aed9-4566-9809-6c9d91131896'] AND status == 'BACKUP_FAILED' on node2 Executing query select dictGet('RedisSimple_hashed_DateTime_', 'DateTime_', toUInt64(2)) on node Executing query DROP TABLE IF EXISTS b on instance run container_id:roottestconfigsubstitutions-gw6-node3-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestconfigsubstitutions-gw6-node3-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select count() from system.tables where database = 'system' and name in ['query_log', 'text_log'] on node2 http://localhost:None "GET /v1.46/containers/e5a667165acdebe03a7248a291501be53e66923bfb42e2902d4dda904ff2e4e7/json HTTP/1.1" 200 None Stdout:8 copy from /var/lib/clickhouse/shadow/*/data/test/tbl to /var/lib/clickhouse/data/test/tbl1/detached run container_id:roottestbackuprestore-gw9-node-1 detach:False nothrow:False cmd: ['bash', '-c', 'cp -r /var/lib/clickhouse/shadow/*/data/test/tbl -T /var/lib/clickhouse/data/test/tbl1/detached'] Command:[docker exec roottestbackuprestore-gw9-node-1 bash -c cp -r /var/lib/clickhouse/shadow/*/data/test/tbl -T /var/lib/clickhouse/data/test/tbl1/detached] Executing query select dictGetInt64OrDefault('LocalClickHouse_complex_key_cache_', 'Int64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt64(-88)) on local_node Executing query ALTER TABLE test.tbl1 ATTACH PARTITION 197001 on node http://localhost:None "GET /v1.46/containers/e5a667165acdebe03a7248a291501be53e66923bfb42e2902d4dda904ff2e4e7/json HTTP/1.1" 200 None [gw8] PASSED test_allow_feature_tier/test.py::test_allow_feature_tier_in_mergetree_settings_with_old_compatibility test_allow_feature_tier/test.py::test_allow_feature_tier_in_user Executing query DROP USER IF EXISTS user_experimental on instance http://localhost:None "GET /v1.46/containers/e5a667165acdebe03a7248a291501be53e66923bfb42e2902d4dda904ff2e4e7/json HTTP/1.1" 200 None Executing query select dictGetDateTime('RedisSimple_hashed_DateTime_', 'DateTime_', toUInt64(2)) on node Executing query rename table system.text_log to system.text_log_10_test on node2 Executing query ALTER TABLE test.tbl1 ATTACH PARTITION 197002 on node Executing query select dictGet('LocalClickHouse_complex_key_cache_', 'UUID_', tuple(toUInt64(1),toString('world'))) on local_node Executing query DROP TABLE tbl ON CLUSTER 'cluster' SYNC on node1 http://localhost:None "GET /v1.46/containers/e5a667165acdebe03a7248a291501be53e66923bfb42e2902d4dda904ff2e4e7/json HTTP/1.1" 200 None Executing query SELECT value FROM system.server_settings where name = 'allow_feature_tier' on instance http://localhost:None "GET /v1.46/containers/e5a667165acdebe03a7248a291501be53e66923bfb42e2902d4dda904ff2e4e7/json HTTP/1.1" 200 None Executing query rename table system.query_log to system.query_log_10_test on node2 Executing query select dictHas('RedisSimple_hashed_DateTime_', toUInt64(2)) on node Executing query SELECT sleep(2) on node http://localhost:None "GET /v1.46/containers/e5a667165acdebe03a7248a291501be53e66923bfb42e2902d4dda904ff2e4e7/json HTTP/1.1" 200 None run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select dictGetUUID('LocalClickHouse_complex_key_cache_', 'UUID_', tuple(toUInt64(1),toString('world'))) on local_node http://localhost:None "GET /v1.46/containers/e5a667165acdebe03a7248a291501be53e66923bfb42e2902d4dda904ff2e4e7/json HTTP/1.1" 200 None Stdout:795 run container_id:roottestallowfeaturetier-gw8-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "sed -i 's/0/1/g' /etc/clickhouse-server/config.d/allow_feature_tier.xml"] Command:[docker exec roottestallowfeaturetier-gw8-instance-1 bash -c sed -i 's/0/1/g' /etc/clickhouse-server/config.d/allow_feature_tier.xml] Executing query SYSTEM RELOAD CONFIG on instance http://localhost:None "GET /v1.46/containers/e5a667165acdebe03a7248a291501be53e66923bfb42e2902d4dda904ff2e4e7/json HTTP/1.1" 200 None run container_id:roottestasyncloaddatabases-gw7-node2-1 detach:False nothrow:True cmd: ['bash', '-c', 'ps -C clickhouse'] Command:[docker exec -u root roottestasyncloaddatabases-gw7-node2-1 bash -c ps -C clickhouse] Executing query select dictGetOrDefault('RedisSimple_hashed_DateTime_', 'DateTime_', toUInt64(9999999), toDateTime('2018-12-30 00:00:00')) on node Executing query RESTORE TABLE tbl FROM Disk('backups', '2') on node1 Stdout: PID TTY TIME CMD Stdout: 8 ? 00:00:17 clickhouse run container_id:roottestasyncloaddatabases-gw7-node2-1 detach:False nothrow:False cmd: ['bash', '-c', 'pkill clickhouse'] Command:[docker exec -u root roottestasyncloaddatabases-gw7-node2-1 bash -c pkill clickhouse] http://localhost:None "GET /v1.46/containers/e5a667165acdebe03a7248a291501be53e66923bfb42e2902d4dda904ff2e4e7/json HTTP/1.1" 200 None run container_id:roottestasyncloaddatabases-gw7-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw7-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:8 http://localhost:None "GET /v1.46/containers/e5a667165acdebe03a7248a291501be53e66923bfb42e2902d4dda904ff2e4e7/json HTTP/1.1" 200 None Executing query select dictHas('LocalClickHouse_complex_key_cache_', tuple(toUInt64(1),toString('world'))) on local_node http://localhost:None "GET /v1.46/containers/e5a667165acdebe03a7248a291501be53e66923bfb42e2902d4dda904ff2e4e7/json HTTP/1.1" 200 None run container_id:roottestconfigsubstitutions-gw6-node3-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestconfigsubstitutions-gw6-node3-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query SELECT value FROM system.server_settings where name = 'allow_feature_tier' on instance Executing query select dictGetDateTimeOrDefault('RedisSimple_hashed_DateTime_', 'DateTime_', toUInt64(9999999), toDateTime('2018-12-30 00:00:00')) on node Stdout:8 http://localhost:None "GET /v1.46/containers/e5a667165acdebe03a7248a291501be53e66923bfb42e2902d4dda904ff2e4e7/json HTTP/1.1" 200 None Executing query SELECT * FROM tbl on node1 http://localhost:None "GET /v1.46/containers/e5a667165acdebe03a7248a291501be53e66923bfb42e2902d4dda904ff2e4e7/json HTTP/1.1" 200 None Executing query select dictGetOrDefault('LocalClickHouse_complex_key_cache_', 'UUID_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUUID('550e8400-0000-0000-0000-000000000000')) on local_node Executing query system reload dictionary RedisSimple_cache_DateTime_ on node http://localhost:None "GET /v1.46/containers/e5a667165acdebe03a7248a291501be53e66923bfb42e2902d4dda904ff2e4e7/json HTTP/1.1" 200 None Executing query CREATE USER user_experimental IDENTIFIED WITH no_password SETTINGS allow_experimental_time_series_table = 1 on instance [gw2] PASSED test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[http-True] Executing query DROP TABLE IF EXISTS tbl ON CLUSTER 'cluster3' SYNC on node1 http://localhost:None "GET /v1.46/containers/e5a667165acdebe03a7248a291501be53e66923bfb42e2902d4dda904ff2e4e7/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/e5a667165acdebe03a7248a291501be53e66923bfb42e2902d4dda904ff2e4e7/json HTTP/1.1" 200 None Executing query select dictGet('RedisSimple_cache_DateTime_', 'DateTime_', toUInt64(1)) on node Executing query select dictGetUUIDOrDefault('LocalClickHouse_complex_key_cache_', 'UUID_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUUID('550e8400-0000-0000-0000-000000000000')) on local_node run container_id:roottestallowfeaturetier-gw8-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "sed -i 's/1/0/g' /etc/clickhouse-server/config.d/allow_feature_tier.xml"] Command:[docker exec roottestallowfeaturetier-gw8-instance-1 bash -c sed -i 's/1/0/g' /etc/clickhouse-server/config.d/allow_feature_tier.xml] run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] http://localhost:None "GET /v1.46/containers/e5a667165acdebe03a7248a291501be53e66923bfb42e2902d4dda904ff2e4e7/json HTTP/1.1" 200 None Executing query SYSTEM RELOAD CONFIG on instance run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] No clickhouse process running. Start new one. http://localhost:None "POST /v1.46/containers/roottestdatabasebackup-gw1-instance-1/exec HTTP/1.1" 201 74 http://localhost:None "GET /v1.46/containers/e5a667165acdebe03a7248a291501be53e66923bfb42e2902d4dda904ff2e4e7/json HTTP/1.1" 200 None ClickHouse node started pymysql connect root, clickhouse, 172.16.10.2, 3306 http://localhost:None "POST /v1.46/exec/d7b6dad0e112de4200c4dbd08ac51ce49c8931a7f8f530b4f9a51e19816b5985/start HTTP/1.1" 200 0 http://localhost:None "GET /v1.46/exec/d7b6dad0e112de4200c4dbd08ac51ce49c8931a7f8f530b4f9a51e19816b5985/json HTTP/1.1" 200 586 pymysql connect root, clickhouse, 172.16.10.2, 3306 pymysql connect root, clickhouse, 172.16.10.2, 3306 Executing query select dictGetDateTime('RedisSimple_cache_DateTime_', 'DateTime_', toUInt64(1)) on node Executing query DROP TABLE IF EXISTS tbl2 ON CLUSTER 'cluster3' SYNC on node1 pymysql connect root, clickhouse, 172.16.10.2, 3306 pymysql connect root, clickhouse, 172.16.10.2, 3306 Executing query select dictGet('LocalClickHouse_complex_key_cache_', 'Date_', tuple(toUInt64(1),toString('world'))) on local_node run container_id:roottestasyncloaddatabases-gw7-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw7-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] pymysql connect root, clickhouse, 172.16.10.2, 3306 Stdout:8 pymysql connect root, clickhouse, 172.16.10.2, 3306 pymysql connect root, clickhouse, 172.16.10.2, 3306 Executing query system reload dictionaries on node Executing query SELECT value FROM system.server_settings where name = 'allow_feature_tier' on instance run container_id:roottestconfigsubstitutions-gw6-node3-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestconfigsubstitutions-gw6-node3-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select dictHas('RedisSimple_cache_DateTime_', toUInt64(1)) on node run container_id:roottestconfigsubstitutions-gw6-node3-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestconfigsubstitutions-gw6-node3-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select dictGetDate('LocalClickHouse_complex_key_cache_', 'Date_', tuple(toUInt64(1),toString('world'))) on local_node No clickhouse process running. Start new one. http://localhost:None "POST /v1.46/containers/roottestconfigsubstitutions-gw6-node3-1/exec HTTP/1.1" 201 74 Executing query DROP DATABASE IF EXISTS mydb ON CLUSTER 'cluster3' SYNC on node1 http://localhost:None "POST /v1.46/exec/4b47dab408fd642a53ea97d66116e6c4a9aa0068e2ee778c9f90e3a82c09630d/start HTTP/1.1" 200 0 Executing query select dictGet('MySQL_complex_key_cache_', 'UInt8_', tuple(toUInt64(1),toString('world'))) on node http://localhost:None "GET /v1.46/exec/4b47dab408fd642a53ea97d66116e6c4a9aa0068e2ee778c9f90e3a82c09630d/json HTTP/1.1" 200 586 Executing query CREATE USER user_experimental IDENTIFIED WITH no_password SETTINGS allow_experimental_time_series_table = 1 on instance Executing query select dictGetOrDefault('RedisSimple_cache_DateTime_', 'DateTime_', toUInt64(9999999), toDateTime('2018-12-30 00:00:00')) on node Executing query select dictHas('LocalClickHouse_complex_key_cache_', tuple(toUInt64(1),toString('world'))) on local_node Executing query SELECT value FROM system.settings WHERE name = 'allow_experimental_time_series_table' on instance Executing query select dictGetUInt8('MySQL_complex_key_cache_', 'UInt8_', tuple(toUInt64(1),toString('world'))) on node Executing query DROP DATABASE IF EXISTS mydb2 ON CLUSTER 'cluster3' SYNC on node1 Executing query select dictGetDateTimeOrDefault('RedisSimple_cache_DateTime_', 'DateTime_', toUInt64(9999999), toDateTime('2018-12-30 00:00:00')) on node Executing query select dictGetOrDefault('LocalClickHouse_complex_key_cache_', 'Date_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDate('2018-12-30')) on local_node run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query SELECT value FROM system.settings WHERE name = 'allow_experimental_time_series_table' on instance Executing query SELECT * FROM test.tbl1 ORDER BY p on node Executing query select dictHas('MySQL_complex_key_cache_', tuple(toUInt64(1),toString('world'))) on node Stdout:1603 Clickhouse process running. run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:1603 Executing query select 20 on instance Executing query select dictGet('RedisSimple_cache_DateTime_', 'DateTime_', toUInt64(2)) on node Executing query DROP USER IF EXISTS u1, u2 ON CLUSTER 'cluster3' on node1 run container_id:roottestasyncloaddatabases-gw7-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw7-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select dictGetDateOrDefault('LocalClickHouse_complex_key_cache_', 'Date_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDate('2018-12-30')) on local_node Stdout:8 Executing query select dictGetOrDefault('MySQL_complex_key_cache_', 'UInt8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt8(55)) on node Executing query ALTER TABLE test.tbl1 UPDATE k=10 WHERE 1 on node run container_id:roottestallowfeaturetier-gw8-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "sed -i 's/0/1/g' /etc/clickhouse-server/config.d/allow_feature_tier.xml"] Command:[docker exec roottestallowfeaturetier-gw8-instance-1 bash -c sed -i 's/0/1/g' /etc/clickhouse-server/config.d/allow_feature_tier.xml] Executing query SYSTEM RELOAD CONFIG on instance Executing query select dictGetDateTime('RedisSimple_cache_DateTime_', 'DateTime_', toUInt64(2)) on node run container_id:roottestconfigsubstitutions-gw6-node3-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestconfigsubstitutions-gw6-node3-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select dictGet('LocalClickHouse_complex_key_cache_', 'DateTime_', tuple(toUInt64(1),toString('world'))) on local_node Stdout:787 Clickhouse process running. run container_id:roottestconfigsubstitutions-gw6-node3-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestconfigsubstitutions-gw6-node3-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[native-True] Executing query CREATE TABLE tbl ON CLUSTER 'cluster' (x UInt8) ENGINE=ReplicatedMergeTree('/clickhouse/tables/tbl/', '{replica}')ORDER BY x on node1 Stdout:787 Executing query select 20 on node3 Executing query SELECT sleep(2) on node Executing query select dictGetUInt8OrDefault('MySQL_complex_key_cache_', 'UInt8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt8(55)) on node Executing query select dictHas('RedisSimple_cache_DateTime_', toUInt64(2)) on node Executing query SELECT value FROM system.server_settings where name = 'allow_feature_tier' on instance Executing query select dictGetDateTime('LocalClickHouse_complex_key_cache_', 'DateTime_', tuple(toUInt64(1),toString('world'))) on local_node Executing query select 20 on instance Executing query select dictGet('MySQL_complex_key_cache_', 'UInt16_', tuple(toUInt64(1),toString('world'))) on node Executing query INSERT INTO tbl VALUES (1) on node1 run container_id:roottestallowfeaturetier-gw8-instance-1 detach:False nothrow:True cmd: ['bash', '-c', 'ps -C clickhouse'] Command:[docker exec -u root roottestallowfeaturetier-gw8-instance-1 bash -c ps -C clickhouse] Executing query select dictHas('LocalClickHouse_complex_key_cache_', tuple(toUInt64(1),toString('world'))) on local_node Executing query select dictGetOrDefault('RedisSimple_cache_DateTime_', 'DateTime_', toUInt64(9999999), toDateTime('2018-12-30 00:00:00')) on node Stdout: PID TTY TIME CMD Stdout: 775 ? 00:00:05 clickhouse run container_id:roottestallowfeaturetier-gw8-instance-1 detach:False nothrow:False cmd: ['bash', '-c', 'pkill clickhouse'] Command:[docker exec -u root roottestallowfeaturetier-gw8-instance-1 bash -c pkill clickhouse] run container_id:roottestallowfeaturetier-gw8-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestallowfeaturetier-gw8-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select dictGetUInt16('MySQL_complex_key_cache_', 'UInt16_', tuple(toUInt64(1),toString('world'))) on node Stdout:775 run container_id:roottestasyncloaddatabases-gw7-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw7-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select 20 on node3 Executing query BACKUP TABLE tbl ON CLUSTER 'cluster' TO Disk('backups', '3') ASYNC on node1 Stdout:8 Executing query select dictGetOrDefault('LocalClickHouse_complex_key_cache_', 'DateTime_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDateTime('2018-12-30 00:00:00')) on local_node Executing query select dictGetDateTimeOrDefault('RedisSimple_cache_DateTime_', 'DateTime_', toUInt64(9999999), toDateTime('2018-12-30 00:00:00')) on node Executing query select dictHas('MySQL_complex_key_cache_', tuple(toUInt64(1),toString('world'))) on node Executing query select 20 on instance Executing query select dictGetDateTimeOrDefault('LocalClickHouse_complex_key_cache_', 'DateTime_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDateTime('2018-12-30 00:00:00')) on local_node Executing query BACKUP TABLE tbl ON CLUSTER 'cluster' TO Disk('backups', '3') ASYNC on node2 Executing query system reload dictionary RedisSimple_direct_DateTime_ on node Executing query select dictGetOrDefault('MySQL_complex_key_cache_', 'UInt16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt16(66)) on node Executing query select dictGet('RedisSimple_direct_DateTime_', 'DateTime_', toUInt64(1)) on node Executing query SELECT status FROM system.backups WHERE id IN ['0ae525f2-09a5-4c1c-b791-3369d129c105', 'ba9f45d9-b8be-40d8-a38a-1f78db778766'] AND status == 'CREATING_BACKUP' on node1 Executing query select dictGet('LocalClickHouse_complex_key_cache_', 'String_', tuple(toUInt64(1),toString('world'))) on local_node Executing query select 20 on node3 Executing query select dictGetUInt16OrDefault('MySQL_complex_key_cache_', 'UInt16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt16(66)) on node run container_id:roottestallowfeaturetier-gw8-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestallowfeaturetier-gw8-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select dictGetDateTime('RedisSimple_direct_DateTime_', 'DateTime_', toUInt64(1)) on node Executing query select 20 on instance Executing query SELECT status FROM system.backups WHERE id IN ['0ae525f2-09a5-4c1c-b791-3369d129c105', 'ba9f45d9-b8be-40d8-a38a-1f78db778766'] AND status == 'CREATING_BACKUP' on node2 Executing query select dictGetString('LocalClickHouse_complex_key_cache_', 'String_', tuple(toUInt64(1),toString('world'))) on local_node Stdout:775 Executing query select dictGet('MySQL_complex_key_cache_', 'UInt32_', tuple(toUInt64(1),toString('world'))) on node run container_id:roottestasyncloaddatabases-gw7-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw7-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:8 Executing query select dictHas('RedisSimple_direct_DateTime_', toUInt64(1)) on node Executing query SELECT count() FROM system.backups WHERE id IN ['0ae525f2-09a5-4c1c-b791-3369d129c105', 'ba9f45d9-b8be-40d8-a38a-1f78db778766'] AND status == 'BACKUP_CREATED' on node1 Executing query SELECT id, value FROM test_database_backup.test_table_1 on instance Executing query select dictHas('LocalClickHouse_complex_key_cache_', tuple(toUInt64(1),toString('world'))) on local_node Executing query select dictGetUInt32('MySQL_complex_key_cache_', 'UInt32_', tuple(toUInt64(1),toString('world'))) on node Executing query select 20 on node3 [gw4] PASSED test_cluster_all_replicas/test.py::test_error_on_unavailable_replica[two_shards_three_nodes] Executing query DROP TABLE IF EXISTS u; on node1 test_cluster_all_replicas/test.py::test_global_in Executing query select dictGetOrDefault('RedisSimple_direct_DateTime_', 'DateTime_', toUInt64(9999999), toDateTime('2018-12-30 00:00:00')) on node Executing query SELECT count() FROM system.backups WHERE id IN ['0ae525f2-09a5-4c1c-b791-3369d129c105', 'ba9f45d9-b8be-40d8-a38a-1f78db778766'] AND status == 'BACKUP_CREATED' on node2 Executing query SELECT id, value FROM test_database_backup.test_table_2 on instance Executing query SELECT * FROM test.tbl1 ORDER BY p on node Executing query select dictGetOrDefault('LocalClickHouse_complex_key_cache_', 'String_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toString('hi')) on local_node Executing query select dictHas('MySQL_complex_key_cache_', tuple(toUInt64(1),toString('world'))) on node Executing query CREATE TABLE u(uid Int16) ENGINE=Memory as select 0 on node1 Executing query SYSTEM RELOAD CONFIG on node3 Executing query select dictGetDateTimeOrDefault('RedisSimple_direct_DateTime_', 'DateTime_', toUInt64(9999999), toDateTime('2018-12-30 00:00:00')) on node Executing query SELECT count() FROM system.backups WHERE id IN ['0ae525f2-09a5-4c1c-b791-3369d129c105', 'ba9f45d9-b8be-40d8-a38a-1f78db778766'] AND status == 'BACKUP_FAILED' on node1 Executing query SELECT id, value FROM test_database_backup.test_table_3 on instance Executing query select dictGetOrDefault('MySQL_complex_key_cache_', 'UInt32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt32(77)) on node Executing query DROP TABLE IF EXISTS test.tbl1 on node Executing query select dictGetStringOrDefault('LocalClickHouse_complex_key_cache_', 'String_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toString('hi')) on local_node Executing query SELECT hostName(), * FROM clusterAllReplicas("one_shard_two_nodes", system.one) where dummy GLOBAL IN u ORDER BY ALL on node1 run container_id:roottestallowfeaturetier-gw8-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestallowfeaturetier-gw8-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:775 Executing query SELECT value FROM system.merge_tree_settings WHERE name='min_bytes_for_wide_part' on node3 Executing query select dictGet('RedisSimple_direct_DateTime_', 'DateTime_', toUInt64(2)) on node Executing query DROP DATABASE test_database_backup on instance Executing query SELECT count() FROM system.backups WHERE id IN ['0ae525f2-09a5-4c1c-b791-3369d129c105', 'ba9f45d9-b8be-40d8-a38a-1f78db778766'] AND status == 'BACKUP_FAILED' on node2 Executing query select dictGetUInt32OrDefault('MySQL_complex_key_cache_', 'UInt32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt32(77)) on node run container_id:roottestasyncloaddatabases-gw7-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw7-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Command:[docker compose --env-file /ClickHouse/tests/integration/test_backup_restore/_instances-0-gw9/.env --project-name roottestbackuprestore-gw9 --file /ClickHouse/tests/integration/test_backup_restore/_instances-0-gw9/node/docker-compose.yml stop --timeout 20] [gw9] PASSED test_backup_restore/test.py::test_restore run container_id:roottestasyncloaddatabases-gw7-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw7-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select dictGet('LocalClickHouse_complex_key_cache_', 'Float32_', tuple(toUInt64(1),toString('world'))) on local_node Executing query SELECT hostName() FROM clusterAllReplicas('one_shard_three_nodes', system.one) ORDER BY ALL settings skip_unavailable_shards=1 on node1 [gw4] PASSED test_cluster_all_replicas/test.py::test_global_in test_cluster_all_replicas/test.py::test_skip_unavailable_replica[one_shard_three_nodes] No clickhouse process running. Start new one. http://localhost:None "POST /v1.46/containers/roottestasyncloaddatabases-gw7-node2-1/exec HTTP/1.1" 201 74 http://localhost:None "POST /v1.46/exec/94440209646d561033c3b12513c4400ac3eff943bc0e8d45b2636948f86a5d7b/start HTTP/1.1" 200 0 http://localhost:None "GET /v1.46/exec/94440209646d561033c3b12513c4400ac3eff943bc0e8d45b2636948f86a5d7b/json HTTP/1.1" 200 586 Executing query DROP DATABASE test_database on instance Executing query select dictGetDateTime('RedisSimple_direct_DateTime_', 'DateTime_', toUInt64(2)) on node Executing query SELECT value FROM system.merge_tree_settings WHERE name='min_rows_for_wide_part' on node3 Executing query DROP TABLE tbl ON CLUSTER 'cluster' SYNC on node1 Executing query select dictGet('MySQL_complex_key_cache_', 'UInt64_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGetFloat32('LocalClickHouse_complex_key_cache_', 'Float32_', tuple(toUInt64(1),toString('world'))) on local_node run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', 'rm -rf /backups/'] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c rm -rf /backups/] run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', 'rm -rf /local_plain/'] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c rm -rf /local_plain/] Executing query select dictHas('RedisSimple_direct_DateTime_', toUInt64(2)) on node http://172.16.8.4:9001 "GET /root?delimiter=&encoding-type=url&list-type=2&max-keys=1000&prefix= HTTP/1.1" 200 0 [gw1] PASSED test_database_backup/test.py::test_database_backup_database[Disk('backup_disk_object_storage_local_plain', 'test_database_backup')] test_database_backup/test.py::test_database_backup_database[Disk('backup_disk_s3_plain', 'test_database_backup')] run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', 'rm -rf /backups/'] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c rm -rf /backups/] Executing query SELECT value FROM system.merge_tree_settings WHERE name='merge_max_block_size' on node3 run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', 'rm -rf /local_plain/'] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c rm -rf /local_plain/] http://172.16.8.4:9001 "GET /root?delimiter=&encoding-type=url&list-type=2&max-keys=1000&prefix= HTTP/1.1" 200 0 Executing query DROP DATABASE IF EXISTS test_database; DROP DATABASE IF EXISTS test_database_backup; CREATE DATABASE test_database; CREATE TABLE test_database.test_table_1 (id UInt64, value String) ENGINE=MergeTree ORDER BY id; INSERT INTO test_database.test_table_1 VALUES (0, 'test_database.test_table_1'); CREATE TABLE test_database.test_table_2 (id UInt64, value String) ENGINE=MergeTree ORDER BY id; INSERT INTO test_database.test_table_2 VALUES (0, 'test_database.test_table_2'); CREATE TABLE test_database.test_table_3 (id UInt64, value String) ENGINE=MergeTree ORDER BY id; INSERT INTO test_database.test_table_3 VALUES (0, 'test_database.test_table_3'); BACKUP DATABASE test_database TO Disk('backup_disk_s3_plain', 'test_database_backup'); CREATE DATABASE test_database_backup ENGINE = Backup('test_database', Disk('backup_disk_s3_plain', 'test_database_backup')); on instance Executing query select dictGetUInt64('MySQL_complex_key_cache_', 'UInt64_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictHas('LocalClickHouse_complex_key_cache_', tuple(toUInt64(1),toString('world'))) on local_node Executing query RESTORE TABLE tbl FROM Disk('backups', '3') on node1 Executing query select dictGetOrDefault('RedisSimple_direct_DateTime_', 'DateTime_', toUInt64(9999999), toDateTime('2018-12-30 00:00:00')) on node Executing query SELECT value FROM system.server_settings WHERE name='background_pool_size' on node3 Executing query select dictHas('MySQL_complex_key_cache_', tuple(toUInt64(1),toString('world'))) on node run container_id:roottestallowfeaturetier-gw8-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestallowfeaturetier-gw8-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select dictGetOrDefault('LocalClickHouse_complex_key_cache_', 'Float32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat32(555.11)) on local_node Stdout:775 Executing query select dictGetDateTimeOrDefault('RedisSimple_direct_DateTime_', 'DateTime_', toUInt64(9999999), toDateTime('2018-12-30 00:00:00')) on node Executing query SELECT * FROM tbl on node1 Executing query select dictGetOrDefault('MySQL_complex_key_cache_', 'UInt64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt64(88)) on node get_instance_ip instance_name=zoo1 http://localhost:None "GET /v1.46/containers/roottestconfigsubstitutions-gw6-zoo1-1/json HTTP/1.1" 200 None get_kazoo_client: zoo1, ip:172.16.7.4, port:2181, use_ssl:False Connecting to 172.16.7.4(172.16.7.4):2181, use_ssl: False Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Zookeeper connection established, state: CONNECTED Sending request(xid=1): Create(path='/background_pool_size', data=b'72', acl=[ACL(perms=31, acl_list=['ALL'], id=Id(scheme='world', id='anyone'))], flags=0) Received response(xid=1): '/background_pool_size' run container_id:roottestconfigsubstitutions-gw6-node3-1 detach:False nothrow:False cmd: ['bash', '-c', 'test ! -f /etc/clickhouse-server/config.d/config_zk_include_test.xml || mv --no-clobber /etc/clickhouse-server/config.d/config_zk_include_test.xml /tmp/config_zk_include_test_960d927d-2da1-4392-84f4-40f13b06b7c9.xml'] Command:[docker exec roottestconfigsubstitutions-gw6-node3-1 bash -c test ! -f /etc/clickhouse-server/config.d/config_zk_include_test.xml || mv --no-clobber /etc/clickhouse-server/config.d/config_zk_include_test.xml /tmp/config_zk_include_test_960d927d-2da1-4392-84f4-40f13b06b7c9.xml] run container_id:roottestasyncloaddatabases-gw7-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw7-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:787 Clickhouse process running. run container_id:roottestasyncloaddatabases-gw7-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw7-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] run container_id:roottestconfigsubstitutions-gw6-node3-1 detach:False nothrow:False cmd: ['bash', '-c', 'echo \'\n \n \n 44\n \n \n 1\n 1111\n \n\n \n \n \' > /etc/clickhouse-server/config.d/config_zk_include_test.xml'] Command:[docker exec roottestconfigsubstitutions-gw6-node3-1 bash -c echo ' 44 1 1111 ' > /etc/clickhouse-server/config.d/config_zk_include_test.xml] Executing query select dictGetFloat32OrDefault('LocalClickHouse_complex_key_cache_', 'Float32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat32(555.11)) on local_node Executing query SELECT id, value FROM test_database_backup.test_table_1 on instance Executing query SYSTEM RELOAD CONFIG on node3 Stdout:787 Executing query select 20 on node2 Executing query system reload dictionary RedisHash_complex_key_hashed_DateTime_ on node Executing query DROP TABLE IF EXISTS tbl ON CLUSTER 'cluster3' SYNC on node1 [gw2] PASSED test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[native-True] Executing query select dictGetUInt64OrDefault('MySQL_complex_key_cache_', 'UInt64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt64(88)) on node Executing query select dictGet('LocalClickHouse_complex_key_cache_', 'Float64_', tuple(toUInt64(1),toString('world'))) on local_node Executing query select dictGet('RedisHash_complex_key_hashed_DateTime_', 'DateTime_', tuple(toUInt64(1),toString('world'))) on node Executing query SELECT id, value FROM test_database_backup.test_table_2 on instance Executing query SELECT value FROM system.server_settings WHERE name='background_pool_size' on node3 Executing query select dictGet('MySQL_complex_key_cache_', 'Int8_', tuple(toUInt64(1),toString('world'))) on node Executing query DROP TABLE IF EXISTS tbl2 ON CLUSTER 'cluster3' SYNC on node1 Executing query select dictGetFloat64('LocalClickHouse_complex_key_cache_', 'Float64_', tuple(toUInt64(1),toString('world'))) on local_node Executing query select dictGetDateTime('RedisHash_complex_key_hashed_DateTime_', 'DateTime_', tuple(toUInt64(1),toString('world'))) on node Executing query SELECT id, value FROM test_database_backup.test_table_3 on instance Executing query select 20 on node2 run container_id:roottestconfigsubstitutions-gw6-node3-1 detach:False nothrow:False cmd: ['bash', '-c', 'test ! -f /tmp/config_zk_include_test_960d927d-2da1-4392-84f4-40f13b06b7c9.xml || mv /tmp/config_zk_include_test_960d927d-2da1-4392-84f4-40f13b06b7c9.xml /etc/clickhouse-server/config.d/config_zk_include_test.xml'] Command:[docker exec roottestconfigsubstitutions-gw6-node3-1 bash -c test ! -f /tmp/config_zk_include_test_960d927d-2da1-4392-84f4-40f13b06b7c9.xml || mv /tmp/config_zk_include_test_960d927d-2da1-4392-84f4-40f13b06b7c9.xml /etc/clickhouse-server/config.d/config_zk_include_test.xml] run container_id:roottestallowfeaturetier-gw8-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestallowfeaturetier-gw8-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select dictGetInt8('MySQL_complex_key_cache_', 'Int8_', tuple(toUInt64(1),toString('world'))) on node Sending request(xid=2): Delete(path='/background_pool_size', version=-1) Received response(xid=2): True [gw6] PASSED test_config_substitutions/test.py::test_config_multiple_zk_substitutions Executing query select 1 on node4 test_config_substitutions/test.py::test_include_config Executing query select dictHas('LocalClickHouse_complex_key_cache_', tuple(toUInt64(1),toString('world'))) on local_node run container_id:roottestallowfeaturetier-gw8-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestallowfeaturetier-gw8-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select dictHas('RedisHash_complex_key_hashed_DateTime_', tuple(toUInt64(1),toString('world'))) on node No clickhouse process running. Start new one. http://localhost:None "POST /v1.46/containers/roottestallowfeaturetier-gw8-instance-1/exec HTTP/1.1" 201 74 http://localhost:None "POST /v1.46/exec/196b0d913024fa5169ea63db772e5c129972135898ed7b3796986fa2806737ea/start HTTP/1.1" 200 0 http://localhost:None "GET /v1.46/exec/196b0d913024fa5169ea63db772e5c129972135898ed7b3796986fa2806737ea/json HTTP/1.1" 200 586 Executing query DROP DATABASE IF EXISTS mydb ON CLUSTER 'cluster3' SYNC on node1 run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:True cmd: ['bash', '-c', 'ps -C clickhouse'] Command:[docker exec -u root roottestdatabasebackup-gw1-instance-1 bash -c ps -C clickhouse] Stdout: PID TTY TIME CMD Stdout: 1603 ? 00:00:04 clickhouse run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', 'pkill clickhouse'] Command:[docker exec -u root roottestdatabasebackup-gw1-instance-1 bash -c pkill clickhouse] Executing query select dictHas('MySQL_complex_key_cache_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGetOrDefault('LocalClickHouse_complex_key_cache_', 'Float64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat64(777.11)) on local_node run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select 1 on node4 Executing query select dictGetOrDefault('RedisHash_complex_key_hashed_DateTime_', 'DateTime_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDateTime('2018-12-30 00:00:00')) on node Stdout:1603 Executing query DROP DATABASE IF EXISTS mydb2 ON CLUSTER 'cluster3' SYNC on node1 Executing query select dictGetOrDefault('MySQL_complex_key_cache_', 'Int8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt8(-55)) on node Executing query select dictGetFloat64OrDefault('LocalClickHouse_complex_key_cache_', 'Float64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat64(777.11)) on local_node Executing query select 1 on node4 Executing query select dictGetDateTimeOrDefault('RedisHash_complex_key_hashed_DateTime_', 'DateTime_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDateTime('2018-12-30 00:00:00')) on node Executing query select 20 on node2 Executing query DROP USER IF EXISTS u1, u2 ON CLUSTER 'cluster3' on node1 Executing query select dictGetInt8OrDefault('MySQL_complex_key_cache_', 'Int8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt8(-55)) on node Executing query select dictGet('LocalClickHouse_complex_key_cache_', 'UInt8_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select dictGet('RedisHash_complex_key_hashed_DateTime_', 'DateTime_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select 1 on node3 Stderr: Container roottestbackuprestore-gw9-node-1 Stopping Stderr: Container roottestbackuprestore-gw9-node-1 Stopped Command:[bash -c [ -f /ClickHouse/tests/integration/test_backup_restore/_instances-0-gw9/node/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_backup_restore/_instances-0-gw9/node/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[docker compose --env-file /ClickHouse/tests/integration/test_backup_restore/_instances-0-gw9/.env --project-name roottestbackuprestore-gw9 --file /ClickHouse/tests/integration/test_backup_restore/_instances-0-gw9/node/docker-compose.yml down --volumes] run container_id:roottestallowfeaturetier-gw8-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestallowfeaturetier-gw8-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select dictGet('MySQL_complex_key_cache_', 'Int16_', tuple(toUInt64(1),toString('world'))) on node test_backup_restore_on_cluster/test.py::test_backup_restore_on_single_replica Executing query CREATE DATABASE mydb ON CLUSTER 'cluster' ENGINE=Replicated('/clickhouse/path/','{shard}','{replica}') on node1 Stdout:1596 Clickhouse process running. run container_id:roottestallowfeaturetier-gw8-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestallowfeaturetier-gw8-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select dictGetDateTime('RedisHash_complex_key_hashed_DateTime_', 'DateTime_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGetUInt8('LocalClickHouse_complex_key_cache_', 'UInt8_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Stdout:1596 Executing query select 20 on instance Executing query select 1 on node3 run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:1603 Executing query select 20 on node2 Executing query select dictHas('RedisHash_complex_key_hashed_DateTime_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictHas('LocalClickHouse_complex_key_cache_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select dictGetInt16('MySQL_complex_key_cache_', 'Int16_', tuple(toUInt64(1),toString('world'))) on node Executing query select 1 on node3 Executing query CREATE TABLE mydb.test (`name` String, `value` UInt32) ENGINE = ReplicatedMergeTree ORDER BY value on node1 Executing query select count() from system.text_log_4_test on node2 Executing query select dictGetOrDefault('RedisHash_complex_key_hashed_DateTime_', 'DateTime_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDateTime('2018-12-30 00:00:00')) on node Executing query select dictGetOrDefault('LocalClickHouse_complex_key_cache_', 'UInt8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt8(55)) on local_node Executing query select dictHas('MySQL_complex_key_cache_', tuple(toUInt64(1),toString('world'))) on node Stderr: Container roottestbackuprestore-gw9-node-1 Stopping Stderr: Container roottestbackuprestore-gw9-node-1 Stopped Stderr: Container roottestbackuprestore-gw9-node-1 Removing Stderr: Container roottestbackuprestore-gw9-node-1 Removed Stderr: Network roottestbackuprestore-gw9_default Removing Stderr: Network roottestbackuprestore-gw9_default Removed Cleanup called Docker networks for project roottestbackuprestore-gw9 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestbackuprestore-gw9 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestbackuprestore-gw9 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestbackuprestore-gw9-.*-1$' --format '{{.ID}}:{{.Names}}'] Executing query select 1 on node8 Unstopped containers: {} No running containers for project: roottestbackuprestore-gw9 Trying to prune unused networks... Trying to prune unused images... Command:[docker image prune -f] Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Stdout:5 Command:[docker volume prune -f] Stdout:Total reclaimed space: 0B Volumes pruned: 5 test_composable_protocols/test.py::test_connections Running tests in /ClickHouse/tests/integration/test_composable_protocols/test.py Cluster start called. is_up=False Executing query select 20 on instance Docker networks for project roottestcomposableprotocols-gw9 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestcomposableprotocols-gw9 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestcomposableprotocols-gw9 are DRIVER VOLUME NAME Cleanup called Docker networks for project roottestcomposableprotocols-gw9 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestcomposableprotocols-gw9 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestcomposableprotocols-gw9 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestcomposableprotocols-gw9-.*-1$' --format '{{.ID}}:{{.Names}}'] Executing query select dictGetDateTimeOrDefault('RedisHash_complex_key_hashed_DateTime_', 'DateTime_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDateTime('2018-12-30 00:00:00')) on node Unstopped containers: {} No running containers for project: roottestcomposableprotocols-gw9 Trying to prune unused networks... Trying to prune unused images... Command:[docker image prune -f] Executing query select count() from system.query_log_7_test on node2 Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Stdout:5 Command:[docker volume prune -f] Executing query select dictGetUInt8OrDefault('LocalClickHouse_complex_key_cache_', 'UInt8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt8(55)) on local_node Executing query select dictGetOrDefault('MySQL_complex_key_cache_', 'Int16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt16(-66)) on node Stdout:Total reclaimed space: 0B Volumes pruned: 5 Setup directory for instance: server Create directory for configuration generated in this helper Create directory for common tests configuration Copy common configuration from helpers Generate and write macros file Copy custom test config files ['/ClickHouse/tests/integration/test_composable_protocols/configs/server.crt', '/ClickHouse/tests/integration/test_composable_protocols/configs/server.key'] to /ClickHouse/tests/integration/test_composable_protocols/_instances-0-gw9/server/configs/config.d Setup database dir /ClickHouse/tests/integration/test_composable_protocols/_instances-0-gw9/server/database Setup logs dir /ClickHouse/tests/integration/test_composable_protocols/_instances-0-gw9/server/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw'} stored in /ClickHouse/tests/integration/test_composable_protocols/_instances-0-gw9/.env Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found Executing query select 1 on node8 http://localhost:None "GET /version HTTP/1.1" 200 826 Command:[docker compose --env-file /ClickHouse/tests/integration/test_composable_protocols/_instances-0-gw9/.env --project-name roottestcomposableprotocols-gw9 --file /ClickHouse/tests/integration/test_composable_protocols/_instances-0-gw9/server/docker-compose.yml pull] Executing query system reload dictionary RedisHash_complex_key_cache_DateTime_ on node run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query INSERT INTO mydb.test VALUES ('abc', 1), ('def', 2) on node1 Executing query select count() from system.text_log_4_test on node2 Stdout:1603 Executing query select dictGet('LocalClickHouse_complex_key_cache_', 'UInt16_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select dictGetInt16OrDefault('MySQL_complex_key_cache_', 'Int16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt16(-66)) on node Executing query select 1 on node8 Executing query select dictGet('RedisHash_complex_key_cache_DateTime_', 'DateTime_', tuple(toUInt64(1),toString('world'))) on node Executing query select 20 on instance Executing query INSERT INTO mydb.test VALUES ('ghi', 3) on node1 Executing query select dictGetUInt16('LocalClickHouse_complex_key_cache_', 'UInt16_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select count() from system.query_log_3_test on node2 Executing query select dictGet('MySQL_complex_key_cache_', 'Int32_', tuple(toUInt64(1),toString('world'))) on node Command:[docker compose --env-file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw6/.env --project-name roottestconfigsubstitutions-gw6 --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw6/node1/docker-compose.yml --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw6/node2/docker-compose.yml --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw6/node3/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw6/node4/docker-compose.yml --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw6/node5/docker-compose.yml --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw6/node6/docker-compose.yml --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw6/node7/docker-compose.yml --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw6/node8/docker-compose.yml stop --timeout 20] [gw6] PASSED test_config_substitutions/test.py::test_include_config Executing query select dictGetDateTime('RedisHash_complex_key_cache_DateTime_', 'DateTime_', tuple(toUInt64(1),toString('world'))) on node Executing query BACKUP DATABASE mydb TO Disk('backups', '4') on node1 Executing query select dictHas('LocalClickHouse_complex_key_cache_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select dictGetInt32('MySQL_complex_key_cache_', 'Int32_', tuple(toUInt64(1),toString('world'))) on node Executing query select count() from system.text_log_8_test on node2 Executing query select dictHas('RedisHash_complex_key_cache_DateTime_', tuple(toUInt64(1),toString('world'))) on node Executing query DROP DATABASE mydb SYNC on node1 Executing query select 20 on instance Executing query select dictHas('MySQL_complex_key_cache_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGetOrDefault('LocalClickHouse_complex_key_cache_', 'UInt16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt16(66)) on local_node Executing query select dictGetOrDefault('RedisHash_complex_key_cache_DateTime_', 'DateTime_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDateTime('2018-12-30 00:00:00')) on node Executing query select count() from system.query_log_8_test on node2 run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:1603 Executing query RESTORE DATABASE mydb FROM Disk('backups', '4') on node1 Executing query SELECT value FROM system.settings WHERE name = 'allow_experimental_time_series_table' on instance Executing query select dictGetDateTimeOrDefault('RedisHash_complex_key_cache_DateTime_', 'DateTime_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDateTime('2018-12-30 00:00:00')) on node Executing query select dictGetOrDefault('MySQL_complex_key_cache_', 'Int32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt32(-77)) on node Executing query select dictGetUInt16OrDefault('LocalClickHouse_complex_key_cache_', 'UInt16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt16(66)) on local_node Executing query select count() from system.asynchronous_loader where job ilike '%_log_%_test' and execution_pool = 'BackgroundLoad' on node2 Executing query SELECT value FROM system.settings WHERE name = 'allow_experimental_time_series_table' on instance Executing query select dictGet('RedisHash_complex_key_cache_DateTime_', 'DateTime_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGet('LocalClickHouse_complex_key_cache_', 'UInt32_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select dictGetInt32OrDefault('MySQL_complex_key_cache_', 'Int32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt32(-77)) on node Executing query RESTORE DATABASE mydb FROM Disk('backups', '4') SETTINGS structure_only=true on node1 Executing query system flush logs on node2 Executing query select dictGetUInt32('LocalClickHouse_complex_key_cache_', 'UInt32_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query SELECT 1 SETTINGS allow_experimental_time_series_table=1 on instance Executing query select dictGetDateTime('RedisHash_complex_key_cache_DateTime_', 'DateTime_', tuple(toUInt64(2),toString('qwerty2'))) on node [gw4] PASSED test_cluster_all_replicas/test.py::test_skip_unavailable_replica[one_shard_three_nodes] test_cluster_all_replicas/test.py::test_skip_unavailable_replica[two_shards_three_nodes] Executing query SELECT hostName() FROM clusterAllReplicas('two_shards_three_nodes', system.one) ORDER BY ALL settings skip_unavailable_shards=1 on node1 Executing query select dictGet('MySQL_complex_key_cache_', 'Int64_', tuple(toUInt64(1),toString('world'))) on node run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select dictHas('LocalClickHouse_complex_key_cache_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Stdout:1603 Executing query SELECT 1 SETTINGS allow_experimental_time_series_table=0 on instance Executing query SYSTEM SYNC REPLICA mydb.test on node1 Executing query select dictHas('RedisHash_complex_key_cache_DateTime_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGetInt64('MySQL_complex_key_cache_', 'Int64_', tuple(toUInt64(1),toString('world'))) on node Executing query SELECT * FROM mydb.test ORDER BY name on node1 Executing query select dictGetOrDefault('LocalClickHouse_complex_key_cache_', 'UInt32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt32(77)) on local_node run container_id:roottestallowfeaturetier-gw8-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "sed -i 's/1/0/g' /etc/clickhouse-server/config.d/allow_feature_tier.xml"] Command:[docker exec roottestallowfeaturetier-gw8-instance-1 bash -c sed -i 's/1/0/g' /etc/clickhouse-server/config.d/allow_feature_tier.xml] Executing query select dictGetOrDefault('RedisHash_complex_key_cache_DateTime_', 'DateTime_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDateTime('2018-12-30 00:00:00')) on node Executing query select dictHas('MySQL_complex_key_cache_', tuple(toUInt64(1),toString('world'))) on node Executing query SYSTEM RELOAD CONFIG on instance Executing query select dictGetUInt32OrDefault('LocalClickHouse_complex_key_cache_', 'UInt32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt32(77)) on local_node Executing query DROP DATABASE mydb SYNC on node1 Executing query select dictGetDateTimeOrDefault('RedisHash_complex_key_cache_DateTime_', 'DateTime_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDateTime('2018-12-30 00:00:00')) on node Executing query select dictGetOrDefault('MySQL_complex_key_cache_', 'Int64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt64(-88)) on node Executing query select count() from system.tables where database = 'system' and name in ['query_log', 'text_log'] on node2 Executing query RESTORE DATABASE mydb FROM Disk('backups', '4') SETTINGS allow_non_empty_tables=true on node1 run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query SELECT value FROM system.server_settings where name = 'allow_feature_tier' on instance Executing query select dictGet('LocalClickHouse_complex_key_cache_', 'UInt64_', tuple(toUInt64(2),toString('qwerty2'))) on local_node run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] No clickhouse process running. Start new one. http://localhost:None "POST /v1.46/containers/roottestdatabasebackup-gw1-instance-1/exec HTTP/1.1" 201 74 http://localhost:None "POST /v1.46/exec/5fdae4f46088851871b01d73823a342890dce6a4a5b202bee4173be8555c1efb/start HTTP/1.1" 200 0 Executing query system reload dictionary RedisHash_complex_key_direct_DateTime_ on node http://localhost:None "GET /v1.46/exec/5fdae4f46088851871b01d73823a342890dce6a4a5b202bee4173be8555c1efb/json HTTP/1.1" 200 586 Executing query select dictGetInt64OrDefault('MySQL_complex_key_cache_', 'Int64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt64(-88)) on node Executing query DROP USER IF EXISTS user_experimental on instance Executing query rename table system.text_log to system.text_log_11_test on node2 Executing query select dictGet('RedisHash_complex_key_direct_DateTime_', 'DateTime_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGetUInt64('LocalClickHouse_complex_key_cache_', 'UInt64_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query SYSTEM SYNC REPLICA mydb.test on node1 Executing query rename table system.query_log to system.query_log_11_test on node2 [gw8] PASSED test_allow_feature_tier/test.py::test_allow_feature_tier_in_user test_allow_feature_tier/test.py::test_it_is_possible_to_enable_experimental_settings_in_default_profile run container_id:roottestallowfeaturetier-gw8-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "sed -i 's/0/2/g' /etc/clickhouse-server/config.d/allow_feature_tier.xml"] Command:[docker exec roottestallowfeaturetier-gw8-instance-1 bash -c sed -i 's/0/2/g' /etc/clickhouse-server/config.d/allow_feature_tier.xml] Executing query select dictGet('MySQL_complex_key_cache_', 'UUID_', tuple(toUInt64(1),toString('world'))) on node run container_id:roottestallowfeaturetier-gw8-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "sed -i 's/allow_experimental_time_series_table>./allow_experimental_time_series_table>1/g' /etc/clickhouse-server/users.d/users.xml"] Command:[docker exec roottestallowfeaturetier-gw8-instance-1 bash -c sed -i 's/allow_experimental_time_series_table>./allow_experimental_time_series_table>1/g' /etc/clickhouse-server/users.d/users.xml] Executing query SYSTEM RELOAD CONFIG on instance Executing query SELECT * FROM mydb.test ORDER BY name on node1 Executing query system flush logs on node2 Executing query select dictHas('LocalClickHouse_complex_key_cache_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select dictGetDateTime('RedisHash_complex_key_direct_DateTime_', 'DateTime_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGetUUID('MySQL_complex_key_cache_', 'UUID_', tuple(toUInt64(1),toString('world'))) on node Executing query SELECT value FROM system.server_settings where name = 'allow_feature_tier' on instance Executing query select dictGetOrDefault('LocalClickHouse_complex_key_cache_', 'UInt64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt64(88)) on local_node Executing query DROP TABLE IF EXISTS tbl ON CLUSTER 'cluster3' SYNC on node1 [gw2] PASSED test_backup_restore_on_cluster/test.py::test_backup_restore_on_single_replica Executing query select dictHas('RedisHash_complex_key_direct_DateTime_', tuple(toUInt64(1),toString('world'))) on node run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select dictHas('MySQL_complex_key_cache_', tuple(toUInt64(1),toString('world'))) on node Stdout:2464 Clickhouse process running. run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:2464 Executing query select 20 on instance Executing query SELECT value FROM system.settings WHERE name = 'allow_experimental_time_series_table' on instance Executing query select dictGetUInt64OrDefault('LocalClickHouse_complex_key_cache_', 'UInt64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt64(88)) on local_node Executing query select dictGetOrDefault('RedisHash_complex_key_direct_DateTime_', 'DateTime_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDateTime('2018-12-30 00:00:00')) on node Executing query select dictGetOrDefault('MySQL_complex_key_cache_', 'UUID_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUUID('550e8400-0000-0000-0000-000000000000')) on node Executing query DROP TABLE IF EXISTS tbl2 ON CLUSTER 'cluster3' SYNC on node1 Executing query SELECT 1 SETTINGS allow_experimental_time_series_table=0 on instance Executing query select count() from system.tables where database = 'system' and name in ['query_log', 'text_log'] on node2 Executing query select dictGet('LocalClickHouse_complex_key_cache_', 'Int8_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select dictGetDateTimeOrDefault('RedisHash_complex_key_direct_DateTime_', 'DateTime_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDateTime('2018-12-30 00:00:00')) on node run container_id:roottestallowfeaturetier-gw8-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "sed -i 's/2/0/g' /etc/clickhouse-server/config.d/allow_feature_tier.xml"] Command:[docker exec roottestallowfeaturetier-gw8-instance-1 bash -c sed -i 's/2/0/g' /etc/clickhouse-server/config.d/allow_feature_tier.xml] Executing query select dictGetUUIDOrDefault('MySQL_complex_key_cache_', 'UUID_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUUID('550e8400-0000-0000-0000-000000000000')) on node run container_id:roottestallowfeaturetier-gw8-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "sed -i 's/allow_experimental_time_series_table>./allow_experimental_time_series_table>0/g' /etc/clickhouse-server/users.d/users.xml"] Command:[docker exec roottestallowfeaturetier-gw8-instance-1 bash -c sed -i 's/allow_experimental_time_series_table>./allow_experimental_time_series_table>0/g' /etc/clickhouse-server/users.d/users.xml] Executing query DROP DATABASE IF EXISTS mydb ON CLUSTER 'cluster3' SYNC on node1 Executing query SYSTEM RELOAD CONFIG on instance Executing query select dictGet('RedisHash_complex_key_direct_DateTime_', 'DateTime_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select 20 on instance Executing query rename table system.text_log to system.text_log_12_test on node2 Executing query select dictGetInt8('LocalClickHouse_complex_key_cache_', 'Int8_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select dictGet('MySQL_complex_key_cache_', 'Date_', tuple(toUInt64(1),toString('world'))) on node Executing query rename table system.query_log to system.query_log_12_test on node2 Executing query select dictGetDateTime('RedisHash_complex_key_direct_DateTime_', 'DateTime_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query SELECT value FROM system.server_settings where name = 'allow_feature_tier' on instance Executing query select dictHas('LocalClickHouse_complex_key_cache_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query DROP DATABASE IF EXISTS mydb2 ON CLUSTER 'cluster3' SYNC on node1 Executing query system flush logs on node2 Executing query select dictGetDate('MySQL_complex_key_cache_', 'Date_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictHas('RedisHash_complex_key_direct_DateTime_', tuple(toUInt64(2),toString('qwerty2'))) on node Command:[docker compose --env-file /ClickHouse/tests/integration/test_allow_feature_tier/_instances-0-gw8/.env --project-name roottestallowfeaturetier-gw8 --file /ClickHouse/tests/integration/test_allow_feature_tier/_instances-0-gw8/instance/docker-compose.yml stop --timeout 20] [gw8] PASSED test_allow_feature_tier/test.py::test_it_is_possible_to_enable_experimental_settings_in_default_profile Executing query select dictGetOrDefault('LocalClickHouse_complex_key_cache_', 'Int8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt8(-55)) on local_node Executing query DROP USER IF EXISTS u1, u2 ON CLUSTER 'cluster3' on node1 Executing query select 20 on instance Executing query select dictHas('MySQL_complex_key_cache_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGetOrDefault('RedisHash_complex_key_direct_DateTime_', 'DateTime_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDateTime('2018-12-30 00:00:00')) on node Executing query select dictGetInt8OrDefault('LocalClickHouse_complex_key_cache_', 'Int8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt8(-55)) on local_node Stderr: Container roottestallowfeaturetier-gw8-instance-1 Stopping Stderr: Container roottestallowfeaturetier-gw8-instance-1 Stopped Command:[bash -c [ -f /ClickHouse/tests/integration/test_allow_feature_tier/_instances-0-gw8/instance/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_allow_feature_tier/_instances-0-gw8/instance/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[docker compose --env-file /ClickHouse/tests/integration/test_allow_feature_tier/_instances-0-gw8/.env --project-name roottestallowfeaturetier-gw8 --file /ClickHouse/tests/integration/test_allow_feature_tier/_instances-0-gw8/instance/docker-compose.yml down --volumes] Executing query select dictGetOrDefault('MySQL_complex_key_cache_', 'Date_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDate('2018-12-30')) on node Executing query CREATE TABLE tbl (`x` UInt8, `y` String) ENGINE = MergeTree ORDER BY x on node1 test_backup_restore_on_cluster/test.py::test_different_tables_on_nodes Executing query select dictGet('LocalClickHouse_complex_key_cache_', 'Int16_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select dictGetDateTimeOrDefault('RedisHash_complex_key_direct_DateTime_', 'DateTime_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDateTime('2018-12-30 00:00:00')) on node Executing query CREATE TABLE tbl (`w` Int64) ENGINE = MergeTree ORDER BY w on node2 Executing query select dictGetDateOrDefault('MySQL_complex_key_cache_', 'Date_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDate('2018-12-30')) on node Executing query select count() from system.tables where database = 'system' and name in ['query_log', 'text_log'] on node2 Executing query select dictGetInt16('LocalClickHouse_complex_key_cache_', 'Int16_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select 20 on instance Executing query system reload dictionaries on node Executing query INSERT INTO tbl VALUES (1, 'Don''t'), (2, 'count'), (3, 'your'), (4, 'chickens') on node1 Executing query select dictGet('MySQL_complex_key_cache_', 'DateTime_', tuple(toUInt64(1),toString('world'))) on node [gw0] PASSED test_dictionaries_redis/test.py::test_redis_dictionaries[DateTime_] test_dictionaries_redis/test.py::test_redis_dictionaries[Date_] Run test with id: 9 Executing query system reload dictionaries on node Executing query select dictHas('LocalClickHouse_complex_key_cache_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Stderr: Container roottestallowfeaturetier-gw8-instance-1 Stopping Stderr: Container roottestallowfeaturetier-gw8-instance-1 Stopped Stderr: Container roottestallowfeaturetier-gw8-instance-1 Removing Stderr: Container roottestallowfeaturetier-gw8-instance-1 Removed Stderr: Network roottestallowfeaturetier-gw8_default Removing Stderr: Network roottestallowfeaturetier-gw8_default Removed Cleanup called Executing query SELECT id, value FROM test_database_backup.test_table_1 on instance Docker networks for project roottestallowfeaturetier-gw8 are NETWORK ID NAME DRIVER SCOPE Executing query rename table system.text_log to system.text_log_13_test on node2 Docker containers for project roottestallowfeaturetier-gw8 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestallowfeaturetier-gw8 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestallowfeaturetier-gw8-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} No running containers for project: roottestallowfeaturetier-gw8 Trying to prune unused networks... Executing query INSERT INTO tbl VALUES (-333), (-222), (-111), (0), (111) on node2 Trying to prune unused images... Command:[docker image prune -f] Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Stdout:5 Command:[docker volume prune -f] Stdout:Total reclaimed space: 0B Volumes pruned: 5 test_config_decryption/test.py::test_successful_decryption_xml Running tests in /ClickHouse/tests/integration/test_config_decryption/test.py Cluster start called. is_up=False Docker networks for project roottestconfigdecryption-gw8 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestconfigdecryption-gw8 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestconfigdecryption-gw8 are DRIVER VOLUME NAME Cleanup called Executing query select dictGetDateTime('MySQL_complex_key_cache_', 'DateTime_', tuple(toUInt64(1),toString('world'))) on node Executing query system reload dictionary RedisSimple_flat_Date_ on node Docker networks for project roottestconfigdecryption-gw8 are NETWORK ID NAME DRIVER SCOPE Executing query select dictGetOrDefault('LocalClickHouse_complex_key_cache_', 'Int16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt16(-66)) on local_node Docker containers for project roottestconfigdecryption-gw8 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestconfigdecryption-gw8 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestconfigdecryption-gw8-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} No running containers for project: roottestconfigdecryption-gw8 Trying to prune unused networks... Executing query rename table system.query_log to system.query_log_13_test on node2 Trying to prune unused images... Command:[docker image prune -f] Executing query SELECT id, value FROM test_database_backup.test_table_2 on instance Connection dropped: socket connection broken Transition to CONNECTING Zookeeper connection lost Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Executing query BACKUP TABLE tbl ON CLUSTER 'cluster' TO Disk('backups', '5') on node1 Stdout:5 Command:[docker volume prune -f] Stdout:Total reclaimed space: 0B Volumes pruned: 5 Setup directory for instance: node1 Create directory for configuration generated in this helper Create directory for common tests configuration Copy common configuration from helpers Generate and write macros file Copy custom test config files ['/ClickHouse/tests/integration/test_config_decryption/configs/config.xml'] to /ClickHouse/tests/integration/test_config_decryption/_instances-0-gw8/node1/configs/config.d Setup database dir /ClickHouse/tests/integration/test_config_decryption/_instances-0-gw8/node1/database Setup logs dir /ClickHouse/tests/integration/test_config_decryption/_instances-0-gw8/node1/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Setup directory for instance: node2 Create directory for configuration generated in this helper Create directory for common tests configuration Copy common configuration from helpers Generate and write macros file Copy custom test config files ['/ClickHouse/tests/integration/test_config_decryption/configs/config.yaml'] to /ClickHouse/tests/integration/test_config_decryption/_instances-0-gw8/node2/configs/config.d Setup database dir /ClickHouse/tests/integration/test_config_decryption/_instances-0-gw8/node2/database Setup logs dir /ClickHouse/tests/integration/test_config_decryption/_instances-0-gw8/node2/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw'} stored in /ClickHouse/tests/integration/test_config_decryption/_instances-0-gw8/.env Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found http://localhost:None "GET /version HTTP/1.1" 200 826 Command:[docker compose --env-file /ClickHouse/tests/integration/test_config_decryption/_instances-0-gw8/.env --project-name roottestconfigdecryption-gw8 --file /ClickHouse/tests/integration/test_config_decryption/_instances-0-gw8/node1/docker-compose.yml --file /ClickHouse/tests/integration/test_config_decryption/_instances-0-gw8/node2/docker-compose.yml pull] Connecting to 172.16.7.4(172.16.7.4):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Executing query select dictGet('RedisSimple_flat_Date_', 'Date_', toUInt64(1)) on node Executing query select dictHas('MySQL_complex_key_cache_', tuple(toUInt64(1),toString('world'))) on node Executing query system flush logs on node2 Executing query select dictGetInt16OrDefault('LocalClickHouse_complex_key_cache_', 'Int16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt16(-66)) on local_node Connecting to 172.16.7.4(172.16.7.4):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Connection dropped: outstanding heartbeat ping not received Transition to CONNECTING Zookeeper connection lost Executing query SELECT id, value FROM test_database_backup.test_table_3 on instance Executing query select dictGetDate('RedisSimple_flat_Date_', 'Date_', toUInt64(1)) on node Connecting to 172.16.7.4(172.16.7.4):2181, use_ssl: False Executing query select dictGetOrDefault('MySQL_complex_key_cache_', 'DateTime_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDateTime('2018-12-30 00:00:00')) on node Executing query DROP TABLE tbl ON CLUSTER 'cluster' SYNC on node1 Executing query select dictGet('LocalClickHouse_complex_key_cache_', 'Int32_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query DROP DATABASE test_database_backup on instance Executing query select dictGetDateTimeOrDefault('MySQL_complex_key_cache_', 'DateTime_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDateTime('2018-12-30 00:00:00')) on node Executing query select dictHas('RedisSimple_flat_Date_', toUInt64(1)) on node Connecting to 172.16.7.4(172.16.7.4):2181, use_ssl: False Executing query DROP DATABASE test_database on instance Executing query select dictGetInt32('LocalClickHouse_complex_key_cache_', 'Int32_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query RESTORE TABLE tbl ON CLUSTER 'cluster' FROM Disk('backups', '5') on node2 Executing query select dictGetOrDefault('RedisSimple_flat_Date_', 'Date_', toUInt64(9999999), toDate('2018-12-30')) on node Executing query select dictGet('MySQL_complex_key_cache_', 'String_', tuple(toUInt64(1),toString('world'))) on node Executing query select count() from system.tables where database = 'system' and name in ['query_log', 'text_log'] on node2 run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', 'rm -rf /backups/'] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c rm -rf /backups/] Executing query select dictHas('LocalClickHouse_complex_key_cache_', tuple(toUInt64(2),toString('qwerty2'))) on local_node run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', 'rm -rf /local_plain/'] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c rm -rf /local_plain/] http://172.16.8.4:9001 "GET /root?delimiter=&encoding-type=url&list-type=2&max-keys=1000&prefix= HTTP/1.1" 200 0 http://172.16.8.4:9001 "DELETE /root/data/disks/disk_s3_plain/test_database_backup/.backup HTTP/1.1" 204 0 http://172.16.8.4:9001 "DELETE /root/data/disks/disk_s3_plain/test_database_backup/data/test_database/test_table_1/all_1_1_0/checksums.txt HTTP/1.1" 204 0 http://172.16.8.4:9001 "DELETE /root/data/disks/disk_s3_plain/test_database_backup/data/test_database/test_table_1/all_1_1_0/columns.txt HTTP/1.1" 204 0 http://172.16.8.4:9001 "DELETE /root/data/disks/disk_s3_plain/test_database_backup/data/test_database/test_table_1/all_1_1_0/count.txt HTTP/1.1" 204 0 http://172.16.8.4:9001 "DELETE /root/data/disks/disk_s3_plain/test_database_backup/data/test_database/test_table_1/all_1_1_0/data.bin HTTP/1.1" 204 0 Executing query select dictGetDateOrDefault('RedisSimple_flat_Date_', 'Date_', toUInt64(9999999), toDate('2018-12-30')) on node http://172.16.8.4:9001 "DELETE /root/data/disks/disk_s3_plain/test_database_backup/data/test_database/test_table_1/all_1_1_0/data.cmrk3 HTTP/1.1" 204 0 http://172.16.8.4:9001 "DELETE /root/data/disks/disk_s3_plain/test_database_backup/data/test_database/test_table_1/all_1_1_0/default_compression_codec.txt HTTP/1.1" 204 0 http://172.16.8.4:9001 "DELETE /root/data/disks/disk_s3_plain/test_database_backup/data/test_database/test_table_1/all_1_1_0/metadata_version.txt HTTP/1.1" 204 0 http://172.16.8.4:9001 "DELETE /root/data/disks/disk_s3_plain/test_database_backup/data/test_database/test_table_1/all_1_1_0/primary.cidx HTTP/1.1" 204 0 http://172.16.8.4:9001 "DELETE /root/data/disks/disk_s3_plain/test_database_backup/data/test_database/test_table_1/all_1_1_0/serialization.json HTTP/1.1" 204 0 http://172.16.8.4:9001 "DELETE /root/data/disks/disk_s3_plain/test_database_backup/data/test_database/test_table_2/all_1_1_0/checksums.txt HTTP/1.1" 204 0 http://172.16.8.4:9001 "DELETE /root/data/disks/disk_s3_plain/test_database_backup/data/test_database/test_table_2/all_1_1_0/data.bin HTTP/1.1" 204 0 http://172.16.8.4:9001 "DELETE /root/data/disks/disk_s3_plain/test_database_backup/data/test_database/test_table_3/all_1_1_0/checksums.txt HTTP/1.1" 204 0 http://172.16.8.4:9001 "DELETE /root/data/disks/disk_s3_plain/test_database_backup/data/test_database/test_table_3/all_1_1_0/data.bin HTTP/1.1" 204 0 http://172.16.8.4:9001 "DELETE /root/data/disks/disk_s3_plain/test_database_backup/metadata/test_database.sql HTTP/1.1" 204 0 http://172.16.8.4:9001 "DELETE /root/data/disks/disk_s3_plain/test_database_backup/metadata/test_database/test_table_1.sql HTTP/1.1" 204 0 http://172.16.8.4:9001 "DELETE /root/data/disks/disk_s3_plain/test_database_backup/metadata/test_database/test_table_2.sql HTTP/1.1" 204 0 http://172.16.8.4:9001 "DELETE /root/data/disks/disk_s3_plain/test_database_backup/metadata/test_database/test_table_3.sql HTTP/1.1" 204 0 [gw1] PASSED test_database_backup/test.py::test_database_backup_database[Disk('backup_disk_s3_plain', 'test_database_backup')] test_database_backup/test.py::test_database_backup_database[File('test_database_backup_file')] run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', 'rm -rf /backups/'] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c rm -rf /backups/] run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', 'rm -rf /local_plain/'] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c rm -rf /local_plain/] Executing query select dictGetString('MySQL_complex_key_cache_', 'String_', tuple(toUInt64(1),toString('world'))) on node http://172.16.8.4:9001 "GET /root?delimiter=&encoding-type=url&list-type=2&max-keys=1000&prefix= HTTP/1.1" 200 0 Executing query DROP DATABASE IF EXISTS test_database; DROP DATABASE IF EXISTS test_database_backup; CREATE DATABASE test_database; CREATE TABLE test_database.test_table_1 (id UInt64, value String) ENGINE=MergeTree ORDER BY id; INSERT INTO test_database.test_table_1 VALUES (0, 'test_database.test_table_1'); CREATE TABLE test_database.test_table_2 (id UInt64, value String) ENGINE=MergeTree ORDER BY id; INSERT INTO test_database.test_table_2 VALUES (0, 'test_database.test_table_2'); CREATE TABLE test_database.test_table_3 (id UInt64, value String) ENGINE=MergeTree ORDER BY id; INSERT INTO test_database.test_table_3 VALUES (0, 'test_database.test_table_3'); BACKUP DATABASE test_database TO File('test_database_backup_file'); CREATE DATABASE test_database_backup ENGINE = Backup('test_database', File('test_database_backup_file')); on instance Executing query rename table system.text_log to system.text_log_14_test on node2 Executing query select dictGetOrDefault('LocalClickHouse_complex_key_cache_', 'Int32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt32(-77)) on local_node Executing query select dictGet('RedisSimple_flat_Date_', 'Date_', toUInt64(2)) on node Executing query SELECT * FROM tbl on node1 Executing query select dictHas('MySQL_complex_key_cache_', tuple(toUInt64(1),toString('world'))) on node Executing query rename table system.query_log to system.query_log_14_test on node2 Executing query select dictGetInt32OrDefault('LocalClickHouse_complex_key_cache_', 'Int32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt32(-77)) on local_node Executing query SELECT * FROM tbl on node2 Executing query select dictGetDate('RedisSimple_flat_Date_', 'Date_', toUInt64(2)) on node Executing query select dictGetOrDefault('MySQL_complex_key_cache_', 'String_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toString('hi')) on node Executing query system flush logs on node2 Executing query SELECT id, value FROM test_database_backup.test_table_1 on instance Executing query select dictGet('LocalClickHouse_complex_key_cache_', 'Int64_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query DROP TABLE IF EXISTS tbl ON CLUSTER 'cluster3' SYNC on node1 [gw2] PASSED test_backup_restore_on_cluster/test.py::test_different_tables_on_nodes Executing query select dictHas('RedisSimple_flat_Date_', toUInt64(2)) on node Executing query SELECT id, value FROM test_database_backup.test_table_2 on instance Executing query select dictGetStringOrDefault('MySQL_complex_key_cache_', 'String_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toString('hi')) on node Executing query select dictGetInt64('LocalClickHouse_complex_key_cache_', 'Int64_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select dictGetOrDefault('RedisSimple_flat_Date_', 'Date_', toUInt64(9999999), toDate('2018-12-30')) on node Executing query SELECT id, value FROM test_database_backup.test_table_3 on instance Executing query DROP TABLE IF EXISTS tbl2 ON CLUSTER 'cluster3' SYNC on node1 Executing query select dictGet('MySQL_complex_key_cache_', 'Float32_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictHas('LocalClickHouse_complex_key_cache_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select dictGetDateOrDefault('RedisSimple_flat_Date_', 'Date_', toUInt64(9999999), toDate('2018-12-30')) on node run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:True cmd: ['bash', '-c', 'ps -C clickhouse'] Command:[docker exec -u root roottestdatabasebackup-gw1-instance-1 bash -c ps -C clickhouse] Executing query select dictGetFloat32('MySQL_complex_key_cache_', 'Float32_', tuple(toUInt64(1),toString('world'))) on node Stdout: PID TTY TIME CMD Stdout: 2464 ? 00:00:04 clickhouse run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', 'pkill clickhouse'] Command:[docker exec -u root roottestdatabasebackup-gw1-instance-1 bash -c pkill clickhouse] run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query DROP DATABASE IF EXISTS mydb ON CLUSTER 'cluster3' SYNC on node1 Stdout:2464 Executing query select count() from system.tables where database = 'system' and name in ['query_log', 'text_log'] on node2 Executing query system reload dictionary RedisSimple_hashed_Date_ on node Executing query select dictGetOrDefault('LocalClickHouse_complex_key_cache_', 'Int64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt64(-88)) on local_node Executing query select dictHas('MySQL_complex_key_cache_', tuple(toUInt64(1),toString('world'))) on node Stderr: node2 Skipped - Image is already being pulled by node1 Stderr: node1 Pulling Stderr: node1 Pulled ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_config_decryption/_instances-0-gw8/.env --project-name roottestconfigdecryption-gw8 --file /ClickHouse/tests/integration/test_config_decryption/_instances-0-gw8/node1/docker-compose.yml --file /ClickHouse/tests/integration/test_config_decryption/_instances-0-gw8/node2/docker-compose.yml up -d --no-recreate') Command:[docker compose --env-file /ClickHouse/tests/integration/test_config_decryption/_instances-0-gw8/.env --project-name roottestconfigdecryption-gw8 --file /ClickHouse/tests/integration/test_config_decryption/_instances-0-gw8/node1/docker-compose.yml --file /ClickHouse/tests/integration/test_config_decryption/_instances-0-gw8/node2/docker-compose.yml up -d --no-recreate] Stderr: server Pulling Stderr: server Pulled ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_composable_protocols/_instances-0-gw9/.env --project-name roottestcomposableprotocols-gw9 --file /ClickHouse/tests/integration/test_composable_protocols/_instances-0-gw9/server/docker-compose.yml up -d --no-recreate') Command:[docker compose --env-file /ClickHouse/tests/integration/test_composable_protocols/_instances-0-gw9/.env --project-name roottestcomposableprotocols-gw9 --file /ClickHouse/tests/integration/test_composable_protocols/_instances-0-gw9/server/docker-compose.yml up -d --no-recreate] Executing query rename table system.text_log to system.text_log_15_test on node2 Executing query select dictGet('RedisSimple_hashed_Date_', 'Date_', toUInt64(1)) on node Executing query select dictGetInt64OrDefault('LocalClickHouse_complex_key_cache_', 'Int64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt64(-88)) on local_node Executing query DROP DATABASE IF EXISTS mydb2 ON CLUSTER 'cluster3' SYNC on node1 Executing query select dictGetOrDefault('MySQL_complex_key_cache_', 'Float32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat32(555.11)) on node Executing query rename table system.query_log to system.query_log_15_test on node2 Executing query select dictGetDate('RedisSimple_hashed_Date_', 'Date_', toUInt64(1)) on node Executing query select dictGet('LocalClickHouse_complex_key_cache_', 'UUID_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select dictGetFloat32OrDefault('MySQL_complex_key_cache_', 'Float32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat32(555.11)) on node Executing query DROP USER IF EXISTS u1, u2 ON CLUSTER 'cluster3' on node1 Stderr: Network roottestconfigdecryption-gw8_default Creating Stderr: Network roottestconfigdecryption-gw8_default Created Stderr: Container roottestconfigdecryption-gw8-node1-1 Creating Stderr: Container roottestconfigdecryption-gw8-node2-1 Creating Stderr: Container roottestconfigdecryption-gw8-node2-1 Created Stderr: Container roottestconfigdecryption-gw8-node1-1 Created Stderr: Container roottestconfigdecryption-gw8-node1-1 Starting Stderr: Container roottestconfigdecryption-gw8-node2-1 Starting Stderr: Container roottestconfigdecryption-gw8-node2-1 Started Stderr: Container roottestconfigdecryption-gw8-node1-1 Started ClickHouse instance created get_instance_ip instance_name=node1 http://localhost:None "GET /v1.46/containers/roottestconfigdecryption-gw8-node1-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node1 http://localhost:None "GET /v1.46/containers/roottestconfigdecryption-gw8-node1-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node1, ip: 172.16.1.3... http://localhost:None "GET /v1.46/containers/roottestconfigdecryption-gw8-node1-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/7b8493b2800c728997535d9cac9ca5d5eea8ab081be2b36df0dcf06e30624c21/json HTTP/1.1" 200 None Executing query system flush logs on node2 http://localhost:None "GET /v1.46/containers/7b8493b2800c728997535d9cac9ca5d5eea8ab081be2b36df0dcf06e30624c21/json HTTP/1.1" 200 None Executing query select dictHas('RedisSimple_hashed_Date_', toUInt64(1)) on node Stderr: Network roottestcomposableprotocols-gw9_default Creating Stderr: Network roottestcomposableprotocols-gw9_default Created Stderr: Container roottestcomposableprotocols-gw9-server-1 Creating Stderr: Container roottestcomposableprotocols-gw9-server-1 Created Stderr: Container roottestcomposableprotocols-gw9-server-1 Starting Stderr: Container roottestcomposableprotocols-gw9-server-1 Started ClickHouse instance created get_instance_ip instance_name=server http://localhost:None "GET /v1.46/containers/roottestcomposableprotocols-gw9-server-1/json HTTP/1.1" 200 None get_instance_ip instance_name=server http://localhost:None "GET /v1.46/containers/roottestcomposableprotocols-gw9-server-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in server, ip: 172.16.4.2... http://localhost:None "GET /v1.46/containers/roottestcomposableprotocols-gw9-server-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/41d0fbb123ef0db57c93194c0d85aec38f694618162e6067827c0c2b6de61479/json HTTP/1.1" 200 None Executing query select dictGet('MySQL_complex_key_cache_', 'Float64_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGetUUID('LocalClickHouse_complex_key_cache_', 'UUID_', tuple(toUInt64(2),toString('qwerty2'))) on local_node run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Connection dropped: socket connection error: No route to host Connection dropped: socket connection error: No route to host http://localhost:None "GET /v1.46/containers/7b8493b2800c728997535d9cac9ca5d5eea8ab081be2b36df0dcf06e30624c21/json HTTP/1.1" 200 None Stdout:2464 http://localhost:None "GET /v1.46/containers/41d0fbb123ef0db57c93194c0d85aec38f694618162e6067827c0c2b6de61479/json HTTP/1.1" 200 None Executing query CREATE TABLE tbl ON CLUSTER 'cluster' (x UInt8, y String) ENGINE=ReplicatedMergeTree('/clickhouse/tables/tbl/', '{replica}')ORDER BY x on node1 http://localhost:None "GET /v1.46/containers/7b8493b2800c728997535d9cac9ca5d5eea8ab081be2b36df0dcf06e30624c21/json HTTP/1.1" 200 None test_backup_restore_on_cluster/test.py::test_empty_replicated_table http://localhost:None "GET /v1.46/containers/41d0fbb123ef0db57c93194c0d85aec38f694618162e6067827c0c2b6de61479/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/7b8493b2800c728997535d9cac9ca5d5eea8ab081be2b36df0dcf06e30624c21/json HTTP/1.1" 200 None Executing query select dictGetOrDefault('RedisSimple_hashed_Date_', 'Date_', toUInt64(9999999), toDate('2018-12-30')) on node Executing query select dictHas('LocalClickHouse_complex_key_cache_', tuple(toUInt64(2),toString('qwerty2'))) on local_node http://localhost:None "GET /v1.46/containers/41d0fbb123ef0db57c93194c0d85aec38f694618162e6067827c0c2b6de61479/json HTTP/1.1" 200 None Executing query select dictGetFloat64('MySQL_complex_key_cache_', 'Float64_', tuple(toUInt64(1),toString('world'))) on node Connecting to 172.16.7.4(172.16.7.4):2181, use_ssl: False http://localhost:None "GET /v1.46/containers/7b8493b2800c728997535d9cac9ca5d5eea8ab081be2b36df0dcf06e30624c21/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/41d0fbb123ef0db57c93194c0d85aec38f694618162e6067827c0c2b6de61479/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/7b8493b2800c728997535d9cac9ca5d5eea8ab081be2b36df0dcf06e30624c21/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/41d0fbb123ef0db57c93194c0d85aec38f694618162e6067827c0c2b6de61479/json HTTP/1.1" 200 None Executing query select dictGetOrDefault('LocalClickHouse_complex_key_cache_', 'UUID_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUUID('550e8400-0000-0000-0000-000000000000')) on local_node http://localhost:None "GET /v1.46/containers/7b8493b2800c728997535d9cac9ca5d5eea8ab081be2b36df0dcf06e30624c21/json HTTP/1.1" 200 None Executing query select dictHas('MySQL_complex_key_cache_', tuple(toUInt64(1),toString('world'))) on node Executing query BACKUP TABLE tbl ON CLUSTER 'cluster' TO Disk('backups', '6') SETTINGS replica_num=1 on node1 Executing query select dictGetDateOrDefault('RedisSimple_hashed_Date_', 'Date_', toUInt64(9999999), toDate('2018-12-30')) on node http://localhost:None "GET /v1.46/containers/41d0fbb123ef0db57c93194c0d85aec38f694618162e6067827c0c2b6de61479/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/7b8493b2800c728997535d9cac9ca5d5eea8ab081be2b36df0dcf06e30624c21/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/41d0fbb123ef0db57c93194c0d85aec38f694618162e6067827c0c2b6de61479/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/7b8493b2800c728997535d9cac9ca5d5eea8ab081be2b36df0dcf06e30624c21/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/41d0fbb123ef0db57c93194c0d85aec38f694618162e6067827c0c2b6de61479/json HTTP/1.1" 200 None Executing query select dictGetOrDefault('MySQL_complex_key_cache_', 'Float64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat64(777.11)) on node Executing query select dictGet('RedisSimple_hashed_Date_', 'Date_', toUInt64(2)) on node Executing query select dictGetUUIDOrDefault('LocalClickHouse_complex_key_cache_', 'UUID_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUUID('550e8400-0000-0000-0000-000000000000')) on local_node Executing query select count() from system.tables where database = 'system' and name in ['query_log', 'text_log'] on node2 http://localhost:None "GET /v1.46/containers/7b8493b2800c728997535d9cac9ca5d5eea8ab081be2b36df0dcf06e30624c21/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/41d0fbb123ef0db57c93194c0d85aec38f694618162e6067827c0c2b6de61479/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/7b8493b2800c728997535d9cac9ca5d5eea8ab081be2b36df0dcf06e30624c21/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/41d0fbb123ef0db57c93194c0d85aec38f694618162e6067827c0c2b6de61479/json HTTP/1.1" 200 None run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] http://localhost:None "GET /v1.46/containers/7b8493b2800c728997535d9cac9ca5d5eea8ab081be2b36df0dcf06e30624c21/json HTTP/1.1" 200 None Executing query select dictGetFloat64OrDefault('MySQL_complex_key_cache_', 'Float64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat64(777.11)) on node http://localhost:None "GET /v1.46/containers/41d0fbb123ef0db57c93194c0d85aec38f694618162e6067827c0c2b6de61479/json HTTP/1.1" 200 None Stdout:2464 Executing query select dictGet('LocalClickHouse_complex_key_cache_', 'Date_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select dictGetDate('RedisSimple_hashed_Date_', 'Date_', toUInt64(2)) on node Executing query DROP TABLE tbl ON CLUSTER 'cluster' SYNC on node1 http://localhost:None "GET /v1.46/containers/7b8493b2800c728997535d9cac9ca5d5eea8ab081be2b36df0dcf06e30624c21/json HTTP/1.1" 200 None Executing query rename table system.text_log to system.text_log_16_test on node2 http://localhost:None "GET /v1.46/containers/41d0fbb123ef0db57c93194c0d85aec38f694618162e6067827c0c2b6de61479/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/7b8493b2800c728997535d9cac9ca5d5eea8ab081be2b36df0dcf06e30624c21/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/41d0fbb123ef0db57c93194c0d85aec38f694618162e6067827c0c2b6de61479/json HTTP/1.1" 200 None Connecting to 172.16.7.4(172.16.7.4):2181, use_ssl: False Executing query select dictGetDate('LocalClickHouse_complex_key_cache_', 'Date_', tuple(toUInt64(2),toString('qwerty2'))) on local_node http://localhost:None "GET /v1.46/containers/7b8493b2800c728997535d9cac9ca5d5eea8ab081be2b36df0dcf06e30624c21/json HTTP/1.1" 200 None Executing query select dictHas('RedisSimple_hashed_Date_', toUInt64(2)) on node Executing query rename table system.query_log to system.query_log_16_test on node2 Executing query select dictGet('MySQL_complex_key_cache_', 'UInt8_', tuple(toUInt64(2),toString('qwerty2'))) on node http://localhost:None "GET /v1.46/containers/41d0fbb123ef0db57c93194c0d85aec38f694618162e6067827c0c2b6de61479/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/7b8493b2800c728997535d9cac9ca5d5eea8ab081be2b36df0dcf06e30624c21/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/41d0fbb123ef0db57c93194c0d85aec38f694618162e6067827c0c2b6de61479/json HTTP/1.1" 200 None Executing query RESTORE TABLE tbl ON CLUSTER 'cluster' FROM Disk('backups', '6') on node1 http://localhost:None "GET /v1.46/containers/7b8493b2800c728997535d9cac9ca5d5eea8ab081be2b36df0dcf06e30624c21/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/41d0fbb123ef0db57c93194c0d85aec38f694618162e6067827c0c2b6de61479/json HTTP/1.1" 200 None Executing query select dictHas('LocalClickHouse_complex_key_cache_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select dictGetOrDefault('RedisSimple_hashed_Date_', 'Date_', toUInt64(9999999), toDate('2018-12-30')) on node Executing query system flush logs on node2 Executing query select dictGetUInt8('MySQL_complex_key_cache_', 'UInt8_', tuple(toUInt64(2),toString('qwerty2'))) on node http://localhost:None "GET /v1.46/containers/7b8493b2800c728997535d9cac9ca5d5eea8ab081be2b36df0dcf06e30624c21/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/41d0fbb123ef0db57c93194c0d85aec38f694618162e6067827c0c2b6de61479/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/7b8493b2800c728997535d9cac9ca5d5eea8ab081be2b36df0dcf06e30624c21/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/41d0fbb123ef0db57c93194c0d85aec38f694618162e6067827c0c2b6de61479/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/7b8493b2800c728997535d9cac9ca5d5eea8ab081be2b36df0dcf06e30624c21/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/41d0fbb123ef0db57c93194c0d85aec38f694618162e6067827c0c2b6de61479/json HTTP/1.1" 200 None Executing query select dictGetOrDefault('LocalClickHouse_complex_key_cache_', 'Date_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDate('2018-12-30')) on local_node Executing query select dictGetDateOrDefault('RedisSimple_hashed_Date_', 'Date_', toUInt64(9999999), toDate('2018-12-30')) on node Executing query select dictHas('MySQL_complex_key_cache_', tuple(toUInt64(2),toString('qwerty2'))) on node http://localhost:None "GET /v1.46/containers/7b8493b2800c728997535d9cac9ca5d5eea8ab081be2b36df0dcf06e30624c21/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/41d0fbb123ef0db57c93194c0d85aec38f694618162e6067827c0c2b6de61479/json HTTP/1.1" 200 None ClickHouse server started http://localhost:None "GET /v1.46/containers/7b8493b2800c728997535d9cac9ca5d5eea8ab081be2b36df0dcf06e30624c21/json HTTP/1.1" 200 None run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] No clickhouse process running. Start new one. http://localhost:None "POST /v1.46/containers/roottestdatabasebackup-gw1-instance-1/exec HTTP/1.1" 201 74 http://localhost:None "GET /v1.46/containers/7b8493b2800c728997535d9cac9ca5d5eea8ab081be2b36df0dcf06e30624c21/json HTTP/1.1" 200 None ClickHouse node1 started get_instance_ip instance_name=node2 http://localhost:None "GET /v1.46/containers/roottestconfigdecryption-gw8-node2-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node2 http://localhost:None "GET /v1.46/containers/roottestconfigdecryption-gw8-node2-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node2, ip: 172.16.1.2... http://localhost:None "GET /v1.46/containers/roottestconfigdecryption-gw8-node2-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/d1932ead5187a80a1b81f400f60f509eba6d943292a68c4a707811bbeca60f97/json HTTP/1.1" 200 None ClickHouse node2 started Executing query select value from system.server_settings where name ='max_table_size_to_drop' on node1 http://localhost:None "POST /v1.46/exec/44f9213be59726def87f8bf1c90fd60370c02f45a1019dfaa84131af614b4c23/start HTTP/1.1" 200 0 http://localhost:None "GET /v1.46/exec/44f9213be59726def87f8bf1c90fd60370c02f45a1019dfaa84131af614b4c23/json HTTP/1.1" 200 586 Executing query SYSTEM SYNC REPLICA ON CLUSTER 'cluster' tbl on node1 Executing query system reload dictionary RedisSimple_cache_Date_ on node Executing query select dictGetDateOrDefault('LocalClickHouse_complex_key_cache_', 'Date_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDate('2018-12-30')) on local_node Executing query select dictGetOrDefault('MySQL_complex_key_cache_', 'UInt8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt8(55)) on node Executing query select dictGet('RedisSimple_cache_Date_', 'Date_', toUInt64(1)) on node Executing query select value from system.server_settings where name ='max_partition_size_to_drop' on node1 Executing query SELECT * FROM tbl on node1 Executing query select dictGet('LocalClickHouse_complex_key_cache_', 'DateTime_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select dictGetUInt8OrDefault('MySQL_complex_key_cache_', 'UInt8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt8(55)) on node Executing query select dictGetDate('RedisSimple_cache_Date_', 'Date_', toUInt64(1)) on node Executing query select count() from system.tables where database = 'system' and name in ['query_log', 'text_log'] on node2 [gw8] PASSED test_config_decryption/test.py::test_successful_decryption_xml test_config_decryption/test.py::test_successful_decryption_yaml Executing query select value from system.server_settings where name ='max_table_size_to_drop' on node2 Executing query SELECT * FROM tbl on node2 Executing query select dictGetDateTime('LocalClickHouse_complex_key_cache_', 'DateTime_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select dictGet('MySQL_complex_key_cache_', 'UInt16_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictHas('RedisSimple_cache_Date_', toUInt64(1)) on node Executing query rename table system.text_log to system.text_log_17_test on node2 Connection dropped: socket connection error: No route to host Connection dropped: socket connection error: No route to host Executing query DROP TABLE IF EXISTS tbl ON CLUSTER 'cluster3' SYNC on node1 [gw2] PASSED test_backup_restore_on_cluster/test.py::test_empty_replicated_table Executing query select value from system.server_settings where name ='max_partition_size_to_drop' on node2 [gw9] PASSED test_composable_protocols/test.py::test_connections test_composable_protocols/test.py::test_http_proxy_1 [gw9] PASSED test_composable_protocols/test.py::test_http_proxy_1 test_composable_protocols/test.py::test_proxy_1 Executing query select dictHas('LocalClickHouse_complex_key_cache_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select dictGetUInt16('MySQL_complex_key_cache_', 'UInt16_', tuple(toUInt64(2),toString('qwerty2'))) on node run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query rename table system.query_log to system.query_log_17_test on node2 Stdout:3281 Clickhouse process running. run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select dictGetOrDefault('RedisSimple_cache_Date_', 'Date_', toUInt64(9999999), toDate('2018-12-30')) on node Stdout:3281 Executing query select 20 on instance Command:[docker compose --env-file /ClickHouse/tests/integration/test_config_decryption/_instances-0-gw8/.env --project-name roottestconfigdecryption-gw8 --file /ClickHouse/tests/integration/test_config_decryption/_instances-0-gw8/node1/docker-compose.yml --file /ClickHouse/tests/integration/test_config_decryption/_instances-0-gw8/node2/docker-compose.yml stop --timeout 20] [gw8] PASSED test_config_decryption/test.py::test_successful_decryption_yaml Executing query select dictGetOrDefault('LocalClickHouse_complex_key_cache_', 'DateTime_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDateTime('2018-12-30 00:00:00')) on local_node Executing query select dictHas('MySQL_complex_key_cache_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query system flush logs on node2 Executing query DROP TABLE IF EXISTS tbl2 ON CLUSTER 'cluster3' SYNC on node1 Executing query SYSTEM FLUSH LOGS on server Executing query select dictGetDateOrDefault('RedisSimple_cache_Date_', 'Date_', toUInt64(9999999), toDate('2018-12-30')) on node Connecting to 172.16.7.4(172.16.7.4):2181, use_ssl: False Executing query select dictGetDateTimeOrDefault('LocalClickHouse_complex_key_cache_', 'DateTime_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDateTime('2018-12-30 00:00:00')) on local_node Executing query select dictGetOrDefault('MySQL_complex_key_cache_', 'UInt16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt16(66)) on node Executing query select dictGet('RedisSimple_cache_Date_', 'Date_', toUInt64(2)) on node Executing query DROP DATABASE IF EXISTS mydb ON CLUSTER 'cluster3' SYNC on node1 Stderr: Container roottestconfigsubstitutions-gw6-node3-1 Stopping Stderr: Container roottestconfigsubstitutions-gw6-node5-1 Stopping Stderr: Container roottestconfigsubstitutions-gw6-node6-1 Stopping Stderr: Container roottestconfigsubstitutions-gw6-node2-1 Stopping Stderr: Container roottestconfigsubstitutions-gw6-node7-1 Stopping Stderr: Container roottestconfigsubstitutions-gw6-node1-1 Stopping Stderr: Container roottestconfigsubstitutions-gw6-node8-1 Stopping Stderr: Container roottestconfigsubstitutions-gw6-node4-1 Stopping Stderr: Container roottestconfigsubstitutions-gw6-node3-1 Stopped Stderr: Container roottestconfigsubstitutions-gw6-zoo2-1 Stopping Stderr: Container roottestconfigsubstitutions-gw6-zoo3-1 Stopping Stderr: Container roottestconfigsubstitutions-gw6-zoo1-1 Stopping Stderr: Container roottestconfigsubstitutions-gw6-zoo2-1 Stopped Stderr: Container roottestconfigsubstitutions-gw6-zoo3-1 Stopped Stderr: Container roottestconfigsubstitutions-gw6-node4-1 Stopped Stderr: Container roottestconfigsubstitutions-gw6-node1-1 Stopped Stderr: Container roottestconfigsubstitutions-gw6-node5-1 Stopped Stderr: Container roottestconfigsubstitutions-gw6-node6-1 Stopped Stderr: Container roottestconfigsubstitutions-gw6-node7-1 Stopped Stderr: Container roottestconfigsubstitutions-gw6-node2-1 Stopped Stderr: Container roottestconfigsubstitutions-gw6-node8-1 Stopped Stderr: Container roottestconfigsubstitutions-gw6-zoo1-1 Stopped Command:[bash -c [ -f /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw6/node1/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw6/node1/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[bash -c [ -f /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw6/node2/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw6/node2/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[bash -c [ -f /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw6/node3/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw6/node3/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[bash -c [ -f /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw6/node4/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw6/node4/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[bash -c [ -f /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw6/node5/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw6/node5/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Executing query select 20 on instance Command:[bash -c [ -f /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw6/node6/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw6/node6/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[bash -c [ -f /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw6/node7/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw6/node7/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[bash -c [ -f /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw6/node8/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw6/node8/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[docker compose --env-file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw6/.env --project-name roottestconfigsubstitutions-gw6 --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw6/node1/docker-compose.yml --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw6/node2/docker-compose.yml --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw6/node3/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw6/node4/docker-compose.yml --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw6/node5/docker-compose.yml --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw6/node6/docker-compose.yml --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw6/node7/docker-compose.yml --file /ClickHouse/tests/integration/test_config_substitutions/_instances-0-gw6/node8/docker-compose.yml down --volumes] Executing query select dictGetUInt16OrDefault('MySQL_complex_key_cache_', 'UInt16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt16(66)) on node Executing query select dictGet('LocalClickHouse_complex_key_cache_', 'String_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select dictGetDate('RedisSimple_cache_Date_', 'Date_', toUInt64(2)) on node Executing query DROP DATABASE IF EXISTS mydb2 ON CLUSTER 'cluster3' SYNC on node1 Executing query select dictGet('MySQL_complex_key_cache_', 'UInt32_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGetString('LocalClickHouse_complex_key_cache_', 'String_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select dictHas('RedisSimple_cache_Date_', toUInt64(2)) on node Executing query SYSTEM FLUSH LOGS on server Executing query DROP USER IF EXISTS u1, u2 ON CLUSTER 'cluster3' on node1 Executing query select dictGetUInt32('MySQL_complex_key_cache_', 'UInt32_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select count() from system.tables where database = 'system' and name in ['query_log', 'text_log'] on node2 Executing query select 20 on instance Executing query select dictHas('LocalClickHouse_complex_key_cache_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select dictGetOrDefault('RedisSimple_cache_Date_', 'Date_', toUInt64(9999999), toDate('2018-12-30')) on node Stderr: Container roottestconfigsubstitutions-gw6-node5-1 Stopping Stderr: Container roottestconfigsubstitutions-gw6-node1-1 Stopping Stderr: Container roottestconfigsubstitutions-gw6-node4-1 Stopping Stderr: Container roottestconfigsubstitutions-gw6-node8-1 Stopping Stderr: Container roottestconfigsubstitutions-gw6-node3-1 Stopping Stderr: Container roottestconfigsubstitutions-gw6-node5-1 Stopped Stderr: Container roottestconfigsubstitutions-gw6-node5-1 Removing Stderr: Container roottestconfigsubstitutions-gw6-node6-1 Stopping Stderr: Container roottestconfigsubstitutions-gw6-node2-1 Stopping Stderr: Container roottestconfigsubstitutions-gw6-node7-1 Stopping Stderr: Container roottestconfigsubstitutions-gw6-node3-1 Stopped Stderr: Container roottestconfigsubstitutions-gw6-node3-1 Removing Stderr: Container roottestconfigsubstitutions-gw6-node4-1 Stopped Stderr: Container roottestconfigsubstitutions-gw6-node4-1 Removing Stderr: Container roottestconfigsubstitutions-gw6-node2-1 Stopped Stderr: Container roottestconfigsubstitutions-gw6-node2-1 Removing Stderr: Container roottestconfigsubstitutions-gw6-node6-1 Stopped Stderr: Container roottestconfigsubstitutions-gw6-node6-1 Removing Stderr: Container roottestconfigsubstitutions-gw6-node7-1 Stopped Stderr: Container roottestconfigsubstitutions-gw6-node7-1 Removing Stderr: Container roottestconfigsubstitutions-gw6-node1-1 Stopped Stderr: Container roottestconfigsubstitutions-gw6-node1-1 Removing Stderr: Container roottestconfigsubstitutions-gw6-node8-1 Stopped Stderr: Container roottestconfigsubstitutions-gw6-node8-1 Removing Stderr: Container roottestconfigsubstitutions-gw6-node8-1 Removed Stderr: Container roottestconfigsubstitutions-gw6-node1-1 Removed Stderr: Container roottestconfigsubstitutions-gw6-node3-1 Removed Stderr: Container roottestconfigsubstitutions-gw6-zoo3-1 Stopping Stderr: Container roottestconfigsubstitutions-gw6-zoo2-1 Stopping Stderr: Container roottestconfigsubstitutions-gw6-zoo1-1 Stopping Stderr: Container roottestconfigsubstitutions-gw6-zoo1-1 Stopped Stderr: Container roottestconfigsubstitutions-gw6-zoo1-1 Removing Stderr: Container roottestconfigsubstitutions-gw6-zoo2-1 Stopped Stderr: Container roottestconfigsubstitutions-gw6-zoo2-1 Removing Stderr: Container roottestconfigsubstitutions-gw6-zoo3-1 Stopped Stderr: Container roottestconfigsubstitutions-gw6-zoo3-1 Removing Stderr: Container roottestconfigsubstitutions-gw6-node6-1 Removed Stderr: Container roottestconfigsubstitutions-gw6-node7-1 Removed Stderr: Container roottestconfigsubstitutions-gw6-node2-1 Removed Stderr: Container roottestconfigsubstitutions-gw6-node5-1 Removed Stderr: Container roottestconfigsubstitutions-gw6-node4-1 Removed Stderr: Container roottestconfigsubstitutions-gw6-zoo2-1 Removed Stderr: Container roottestconfigsubstitutions-gw6-zoo1-1 Removed Stderr: Container roottestconfigsubstitutions-gw6-zoo3-1 Removed Stderr: Network roottestconfigsubstitutions-gw6_default Removing Stderr: Network roottestconfigsubstitutions-gw6_default Removed Cleanup called Docker networks for project roottestconfigsubstitutions-gw6 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestconfigsubstitutions-gw6 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestconfigsubstitutions-gw6 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestconfigsubstitutions-gw6-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} No running containers for project: roottestconfigsubstitutions-gw6 Trying to prune unused networks... Trying to prune unused images... Command:[docker image prune -f] Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Stdout:5 Command:[docker volume prune -f] Stdout:Total reclaimed space: 0B Volumes pruned: 5 Running tests in /ClickHouse/tests/integration/test_attach_table_normalizer/test.py Cluster start called. is_up=False test_attach_table_normalizer/test.py::test_attach_substr Docker networks for project roottestattachtablenormalizer-gw6 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestattachtablenormalizer-gw6 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Executing query select dictHas('MySQL_complex_key_cache_', tuple(toUInt64(2),toString('qwerty2'))) on node Docker volumes for project roottestattachtablenormalizer-gw6 are DRIVER VOLUME NAME Cleanup called test_backup_restore_on_cluster/test.py::test_file_deduplication Executing query rename table system.text_log to system.text_log_18_test on node2 Executing query CREATE TABLE tbl ON CLUSTER 'cluster' ( kKIdhlkXxt Int32 ) ENGINE=ReplicatedMergeTree('/clickhouse/tables/tbl/', '{replica}') ORDER BY tuple() SETTINGS min_bytes_for_wide_part=0 on node1 Executing query select dictGetDateOrDefault('RedisSimple_cache_Date_', 'Date_', toUInt64(9999999), toDate('2018-12-30')) on node Executing query select dictGetOrDefault('LocalClickHouse_complex_key_cache_', 'String_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toString('hi')) on local_node Docker networks for project roottestattachtablenormalizer-gw6 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestattachtablenormalizer-gw6 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestattachtablenormalizer-gw6 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestattachtablenormalizer-gw6-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} No running containers for project: roottestattachtablenormalizer-gw6 Trying to prune unused networks... Trying to prune unused images... Command:[docker image prune -f] Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Stdout:5 Command:[docker volume prune -f] Stdout:Total reclaimed space: 0B Volumes pruned: 5 Setup directory for instance: node Create directory for configuration generated in this helper Create directory for common tests configuration Copy common configuration from helpers Generate and write macros file Copy custom test config files [] to /ClickHouse/tests/integration/test_attach_table_normalizer/_instances-0-gw6/node/configs/config.d Setup database dir /ClickHouse/tests/integration/test_attach_table_normalizer/_instances-0-gw6/node/database Setup logs dir /ClickHouse/tests/integration/test_attach_table_normalizer/_instances-0-gw6/node/logs Entrypoint cmd: bash -c "trap 'pkill tail' INT TERM; clickhouse server --config-file=/etc/clickhouse-server/config.xml --log-file=/var/log/clickhouse-server/clickhouse-server.log --errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log --daemon -- ; coproc tail -f /dev/null; wait $$!" Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw'} stored in /ClickHouse/tests/integration/test_attach_table_normalizer/_instances-0-gw6/.env Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found http://localhost:None "GET /version HTTP/1.1" 200 826 Command:[docker compose --env-file /ClickHouse/tests/integration/test_attach_table_normalizer/_instances-0-gw6/.env --project-name roottestattachtablenormalizer-gw6 --file /ClickHouse/tests/integration/test_attach_table_normalizer/_instances-0-gw6/node/docker-compose.yml pull] Executing query rename table system.query_log to system.query_log_18_test on node2 Executing query select dictGetStringOrDefault('LocalClickHouse_complex_key_cache_', 'String_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toString('hi')) on local_node Executing query select dictGetOrDefault('MySQL_complex_key_cache_', 'UInt32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt32(77)) on node Executing query system reload dictionary RedisSimple_direct_Date_ on node Executing query select 20 on instance Command:[docker compose --env-file /ClickHouse/tests/integration/test_composable_protocols/_instances-0-gw9/.env --project-name roottestcomposableprotocols-gw9 --file /ClickHouse/tests/integration/test_composable_protocols/_instances-0-gw9/server/docker-compose.yml stop --timeout 20] [gw9] PASSED test_composable_protocols/test.py::test_proxy_1 Executing query CREATE TABLE tbl2 ON CLUSTER 'cluster' ( kKIdhlkXxt Int32 ) ENGINE=ReplicatedMergeTree('/clickhouse/tables/tbl/', '{replica}-2') ORDER BY tuple() SETTINGS min_bytes_for_wide_part=0 on node1 Executing query system flush logs on node2 Executing query select dictGet('LocalClickHouse_complex_key_cache_', 'Float32_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select dictGetUInt32OrDefault('MySQL_complex_key_cache_', 'UInt32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt32(77)) on node Executing query select dictGet('RedisSimple_direct_Date_', 'Date_', toUInt64(1)) on node Executing query SELECT id, value FROM test_database_backup.test_table_1 on instance Executing query select dictGetFloat32('LocalClickHouse_complex_key_cache_', 'Float32_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select dictGet('MySQL_complex_key_cache_', 'UInt64_', tuple(toUInt64(2),toString('qwerty2'))) on node Connecting to 172.16.7.4(172.16.7.4):2181, use_ssl: False Executing query INSERT INTO tbl VALUES (3556), (1177), (4004), (4264), (3729), (1438), (2158), (2684), (415), (1917) on node1 Executing query select dictGetDate('RedisSimple_direct_Date_', 'Date_', toUInt64(1)) on node Executing query SELECT id, value FROM test_database_backup.test_table_2 on instance Executing query select dictHas('LocalClickHouse_complex_key_cache_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select dictGetUInt64('MySQL_complex_key_cache_', 'UInt64_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query SYSTEM SYNC REPLICA ON CLUSTER 'cluster' tbl on node1 Executing query select dictHas('RedisSimple_direct_Date_', toUInt64(1)) on node Executing query SELECT id, value FROM test_database_backup.test_table_3 on instance Executing query select dictGetOrDefault('LocalClickHouse_complex_key_cache_', 'Float32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat32(555.11)) on local_node Executing query select dictHas('MySQL_complex_key_cache_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGetOrDefault('RedisSimple_direct_Date_', 'Date_', toUInt64(9999999), toDate('2018-12-30')) on node Executing query select count() from system.tables where database = 'system' and name in ['query_log', 'text_log'] on node2 Executing query DROP DATABASE test_database_backup on instance Executing query SYSTEM SYNC REPLICA ON CLUSTER 'cluster' tbl2 on node1 Stderr: Container roottestconfigdecryption-gw8-node2-1 Stopping Stderr: Container roottestconfigdecryption-gw8-node1-1 Stopping Stderr: Container roottestconfigdecryption-gw8-node1-1 Stopped Stderr: Container roottestconfigdecryption-gw8-node2-1 Stopped Command:[bash -c [ -f /ClickHouse/tests/integration/test_config_decryption/_instances-0-gw8/node1/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_config_decryption/_instances-0-gw8/node1/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[bash -c [ -f /ClickHouse/tests/integration/test_config_decryption/_instances-0-gw8/node2/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_config_decryption/_instances-0-gw8/node2/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[docker compose --env-file /ClickHouse/tests/integration/test_config_decryption/_instances-0-gw8/.env --project-name roottestconfigdecryption-gw8 --file /ClickHouse/tests/integration/test_config_decryption/_instances-0-gw8/node1/docker-compose.yml --file /ClickHouse/tests/integration/test_config_decryption/_instances-0-gw8/node2/docker-compose.yml down --volumes] Executing query select dictGetOrDefault('MySQL_complex_key_cache_', 'UInt64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt64(88)) on node Executing query DROP DATABASE test_database on instance Executing query select dictGetDateOrDefault('RedisSimple_direct_Date_', 'Date_', toUInt64(9999999), toDate('2018-12-30')) on node Executing query select dictGetFloat32OrDefault('LocalClickHouse_complex_key_cache_', 'Float32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat32(555.11)) on local_node Executing query rename table system.text_log to system.text_log_19_test on node2 Executing query BACKUP TABLE tbl, TABLE tbl2 ON CLUSTER 'cluster' TO Disk('backups', '7') on node1 Stderr: Container roottestcomposableprotocols-gw9-server-1 Stopping Stderr: Container roottestcomposableprotocols-gw9-server-1 Stopped Command:[bash -c [ -f /ClickHouse/tests/integration/test_composable_protocols/_instances-0-gw9/server/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_composable_protocols/_instances-0-gw9/server/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[docker compose --env-file /ClickHouse/tests/integration/test_composable_protocols/_instances-0-gw9/.env --project-name roottestcomposableprotocols-gw9 --file /ClickHouse/tests/integration/test_composable_protocols/_instances-0-gw9/server/docker-compose.yml down --volumes] run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', 'rm -rf /backups/'] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c rm -rf /backups/] Executing query rename table system.query_log to system.query_log_19_test on node2 Executing query select dictGetUInt64OrDefault('MySQL_complex_key_cache_', 'UInt64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt64(88)) on node run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', 'rm -rf /local_plain/'] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c rm -rf /local_plain/] Executing query select dictGet('RedisSimple_direct_Date_', 'Date_', toUInt64(2)) on node Executing query select dictGet('LocalClickHouse_complex_key_cache_', 'Float64_', tuple(toUInt64(2),toString('qwerty2'))) on local_node http://172.16.8.4:9001 "GET /root?delimiter=&encoding-type=url&list-type=2&max-keys=1000&prefix= HTTP/1.1" 200 0 [gw1] PASSED test_database_backup/test.py::test_database_backup_database[File('test_database_backup_file')] test_database_backup/test.py::test_database_backup_table[Disk('backup_disk_local', 'test_table_backup')] run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', 'rm -rf /backups/'] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c rm -rf /backups/] run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', 'rm -rf /local_plain/'] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c rm -rf /local_plain/] http://172.16.8.4:9001 "GET /root?delimiter=&encoding-type=url&list-type=2&max-keys=1000&prefix= HTTP/1.1" 200 0 Executing query DROP DATABASE IF EXISTS test_database; DROP DATABASE IF EXISTS test_table_backup; CREATE DATABASE test_database; CREATE TABLE test_database.test_table (id UInt64, value String) ENGINE=MergeTree ORDER BY id; INSERT INTO test_database.test_table VALUES (0, 'test_database.test_table'); BACKUP TABLE test_database.test_table TO Disk('backup_disk_local', 'test_table_backup'); CREATE DATABASE test_table_backup ENGINE = Backup('test_database', Disk('backup_disk_local', 'test_table_backup')); on instance Stderr: Container roottestconfigdecryption-gw8-node1-1 Stopping Stderr: Container roottestconfigdecryption-gw8-node2-1 Stopping Stderr: Container roottestconfigdecryption-gw8-node2-1 Stopped Stderr: Container roottestconfigdecryption-gw8-node2-1 Removing Stderr: Container roottestconfigdecryption-gw8-node1-1 Stopped Stderr: Container roottestconfigdecryption-gw8-node1-1 Removing Stderr: Container roottestconfigdecryption-gw8-node2-1 Removed Stderr: Container roottestconfigdecryption-gw8-node1-1 Removed Stderr: Network roottestconfigdecryption-gw8_default Removing Stderr: Network roottestconfigdecryption-gw8_default Removed Cleanup called Docker networks for project roottestconfigdecryption-gw8 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestconfigdecryption-gw8 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestconfigdecryption-gw8 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestconfigdecryption-gw8-.*-1$' --format '{{.ID}}:{{.Names}}'] Executing query system flush logs on node2 Unstopped containers: {} No running containers for project: roottestconfigdecryption-gw8 Trying to prune unused networks... Trying to prune unused images... Command:[docker image prune -f] Executing query select dictGetFloat64('LocalClickHouse_complex_key_cache_', 'Float64_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Executing query select dictGet('MySQL_complex_key_cache_', 'Int8_', tuple(toUInt64(2),toString('qwerty2'))) on node Stdout:5 Command:[docker volume prune -f] Executing query select dictGetDate('RedisSimple_direct_Date_', 'Date_', toUInt64(2)) on node Stdout:Total reclaimed space: 0B Volumes pruned: 5 ENV DOCKER_KERBEROS_KDC_TAG 9391ecdee8d7 test_azure_blob_storage_zero_copy_replication/test.py::test_zero_copy_replication ENV CLICKHOUSE_TESTS_SERVER_BIN_PATH /clickhouse ENV MSAN_OPTIONS abort_on_error=1 poison_in_dtor=1 ENV JAVA_TOOL_OPTIONS -Djdk.attach.allowAttachSelf=true ENV TSAN_OPTIONS halt_on_error=1 abort_on_error=1 history_size=7 memory_limit_mb=46080 second_deadlock_stack=1 ENV HOSTNAME 1dd346380e80 ENV SHLVL 0 ENV HOME /root ENV OLDPWD / ENV DOCKER_HELPER_TAG 5dc43a6382f0 ENV PYTHONUNBUFFERED 1 ENV DOCKER_PYTHON_BOTTLE_TAG caad4729259e ENV UBSAN_OPTIONS print_stacktrace=1 ENV PYTEST_ADDOPTS --dist=loadfile -n 10 -rfEps --run-id=0 --color=no --durations=0 test_aggregation_memory_efficient/test.py::test_remote test_allow_feature_tier/test.py::test_allow_feature_tier_in_general_settings test_allow_feature_tier/test.py::test_allow_feature_tier_in_mergetree_settings test_allow_feature_tier/test.py::test_allow_feature_tier_in_mergetree_settings_with_old_compatibility test_allow_feature_tier/test.py::test_allow_feature_tier_in_user test_allow_feature_tier/test.py::test_it_is_possible_to_enable_experimental_settings_in_default_profile test_analyzer_compatibility/test.py::test_two_new_versions test_async_load_databases/test.py::test_async_load_system_database test_async_load_databases/test.py::test_dependent_tables test_async_load_databases/test.py::test_dict_get_data test_async_load_databases/test.py::test_multiple_tables test_attach_table_normalizer/test.py::test_attach_substr test_attach_table_normalizer/test.py::test_attach_substr_restart test_azure_blob_storage_zero_copy_replication/test.py::test_zero_copy_replication test_backup_restore/test.py::test_attach_partition test_backup_restore/test.py::test_replace_partition test_backup_restore/test.py::test_restore 'test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[http-False]' 'test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[http-True]' 'test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[native-True]' test_backup_restore_on_cluster/test.py::test_backup_restore_on_single_replica test_backup_restore_on_cluster/test.py::test_different_tables_on_nodes test_backup_restore_on_cluster/test.py::test_empty_replicated_table test_backup_restore_on_cluster/test.py::test_file_deduplication test_backup_restore_on_cluster/test.py::test_get_error_from_other_host test_backup_restore_on_cluster/test.py::test_keeper_value_max_size test_backup_restore_on_cluster/test.py::test_mutation test_backup_restore_on_cluster/test.py::test_projection test_backup_restore_on_cluster/test.py::test_replicated_database test_backup_restore_on_cluster/test.py::test_replicated_database_async test_backup_restore_on_cluster/test.py::test_replicated_database_compare_parts 'test_backup_restore_on_cluster/test.py::test_replicated_database_with_special_macro_in_zk_path[database]' 'test_backup_restore_on_cluster/test.py::test_replicated_database_with_special_macro_in_zk_path[uuid]' test_backup_restore_on_cluster/test.py::test_replicated_table test_backup_restore_on_cluster/test.py::test_replicated_table_after_alters test_backup_restore_on_cluster/test.py::test_replicated_table_restored_into_bigger_cluster test_backup_restore_on_cluster/test.py::test_replicated_table_restored_into_smaller_cluster test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_def test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_insert test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_merge test_backup_restore_on_cluster/test.py::test_replicated_table_with_uuid_in_zkpath test_backup_restore_on_cluster/test.py::test_required_privileges test_backup_restore_on_cluster/test.py::test_shutdown_waits_for_backup test_backup_restore_on_cluster/test.py::test_system_functions test_backup_restore_on_cluster/test.py::test_system_users test_backup_restore_on_cluster/test.py::test_table_in_replicated_database_with_not_synced_def test_backup_restore_on_cluster/test.py::test_table_with_parts_in_queue_considered_non_empty test_backup_restore_on_cluster/test.py::test_tables_dependency test_cluster_all_replicas/test.py::test_cluster 'test_cluster_all_replicas/test.py::test_error_on_unavailable_replica[one_shard_three_nodes]' 'test_cluster_all_replicas/test.py::test_error_on_unavailable_replica[two_shards_three_nodes]' test_cluster_all_replicas/test.py::test_global_in 'test_cluster_all_replicas/test.py::test_skip_unavailable_replica[one_shard_three_nodes]' 'test_cluster_all_replicas/test.py::test_skip_unavailable_replica[two_shards_three_nodes]' test_composable_protocols/test.py::test_connections test_composable_protocols/test.py::test_http_proxy_1 test_composable_protocols/test.py::test_proxy_1 test_config_corresponding_root/test.py::test_work test_config_decryption/test.py::test_successful_decryption_xml test_config_decryption/test.py::test_successful_decryption_yaml test_config_hide_in_preprocessed/test.py::test_hide_in_preprocessed test_config_substitutions/test.py::test_allow_databases test_config_substitutions/test.py::test_config test_config_substitutions/test.py::test_config_from_env_overrides test_config_substitutions/test.py::test_config_merge_from_env_overrides test_config_substitutions/test.py::test_config_multiple_zk_substitutions test_config_substitutions/test.py::test_include_config 'test_database_backup/test.py::test_database_backup_database[Disk('"'"'backup_disk_local'"'"', '"'"'test_database_backup'"'"')]' 'test_database_backup/test.py::test_database_backup_database[Disk('"'"'backup_disk_object_storage_local_plain'"'"', '"'"'test_database_backup'"'"')]' 'test_database_backup/test.py::test_database_backup_database[Disk('"'"'backup_disk_s3_plain'"'"', '"'"'test_database_backup'"'"')]' 'test_database_backup/test.py::test_database_backup_database[File('"'"'test_database_backup_file'"'"')]' 'test_database_backup/test.py::test_database_backup_table[Disk('"'"'backup_disk_local'"'"', '"'"'test_table_backup'"'"')]' 'test_database_backup/test.py::test_database_backup_table[Disk('"'"'backup_disk_object_storage_local_plain'"'"', '"'"'test_table_backup'"'"')]' 'test_database_backup/test.py::test_database_backup_table[Disk('"'"'backup_disk_s3_plain'"'"', '"'"'test_table_backup'"'"')]' 'test_database_backup/test.py::test_database_backup_table[File('"'"'test_table_backup_file'"'"')]' 'test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_complex[complex_key_cache]' 'test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_complex[complex_key_direct]' 'test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_complex[complex_key_hashed]' 'test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_ranged[range_hashed]' 'test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_simple[cache]' 'test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_simple[direct]' 'test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_simple[flat]' 'test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_simple[hashed]' 'test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_complex[complex_key_cache]' 'test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_complex[complex_key_direct]' 'test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_complex[complex_key_hashed]' 'test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_ranged[range_hashed]' 'test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_simple[cache]' 'test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_simple[direct]' 'test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_simple[flat]' 'test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_simple[hashed]' 'test_dictionaries_redis/test.py::test_redis_dictionaries[DateTime_]' 'test_dictionaries_redis/test.py::test_redis_dictionaries[Date_]' 'test_dictionaries_redis/test.py::test_redis_dictionaries[Float32_]' 'test_dictionaries_redis/test.py::test_redis_dictionaries[Float64_]' 'test_dictionaries_redis/test.py::test_redis_dictionaries[Int16_]' 'test_dictionaries_redis/test.py::test_redis_dictionaries[Int32_]' 'test_dictionaries_redis/test.py::test_redis_dictionaries[Int64_]' 'test_dictionaries_redis/test.py::test_redis_dictionaries[Int8_]' 'test_dictionaries_redis/test.py::test_redis_dictionaries[String_]' -vvv -ss ENV CLICKHOUSE_LIBRARY_BRIDGE_BINARY_PATH /clickhouse-library-bridge ENV COMPOSE_HTTP_TIMEOUT 600 ENV DOCKER_MYSQL_PHP_CLIENT_TAG 88be89c1e3b6 ENV DOCKER_DOTNET_CLIENT_TAG 11de0b29a15d ENV CLICKHOUSE_TESTS_CLIENT_BIN_PATH /clickhouse ENV DOCKER_MYSQL_JS_CLIENT_TAG 41ba7c2ec2a1 ENV PATH /spark-3.3.2-bin-hadoop3/bin:/opt/gdb/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin ENV DOCKER_KERBERIZED_HADOOP_TAG latest ENV DOCKER_CHANNEL stable ENV DOCKER_CLIENT_TIMEOUT 300 ENV DOCKER_POSTGRESQL_JAVA_CLIENT_TAG a4eff5c7f4d6 ENV DOCKER_NGINX_DAV_TAG b55ac9cd7519 ENV DOCKER_MYSQL_GOLANG_CLIENT_TAG 9bec2a638e6e ENV PWD /ClickHouse/tests/integration ENV DOCKER_MYSQL_JAVA_CLIENT_TAG 766bff31cfe4 ENV CLICKHOUSE_ODBC_BRIDGE_BINARY_PATH /clickhouse-odbc-bridge ENV CLICKHOUSE_TESTS_BASE_CONFIG_DIR /clickhouse-config ENV TZ Etc/UTC ENV JAVA_PATH /usr/lib/jvm/java-11-openjdk-amd64/bin/java ENV DOCKER_BASE_TAG 8b2301119731 ENV SPARK_HOME /spark-3.3.2-bin-hadoop3 ENV LC_CTYPE C.UTF-8 ENV INTEGRATION_TESTS_RUN_ID 0 ENV WORKER_FREE_PORTS 30400 30401 30402 30403 30404 30405 30406 30407 30408 30409 30410 30411 30412 30413 30414 30415 30416 30417 30418 30419 30420 30421 30422 30423 30424 30425 30426 30427 30428 30429 30430 30431 30432 30433 30434 30435 30436 30437 30438 30439 30440 30441 30442 30443 30444 30445 30446 30447 30448 30449 ENV PYTEST_XDIST_TESTRUNUID 9b1b3f7145bb4931b6ac2b8301012c9f ENV PYTEST_XDIST_WORKER gw8 ENV PYTEST_XDIST_WORKER_COUNT 10 ENV PYTEST_CURRENT_TEST test_azure_blob_storage_zero_copy_replication/test.py::test_zero_copy_replication (setup) CLUSTER INIT base_config_dir:/clickhouse-config clickhouse_start_command: clickhouse server --config-file=/etc/clickhouse-server/{main_config_file} --log-file=/var/log/clickhouse-server/clickhouse-server.log --errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log Setup Keeper Cluster name: project_name:roottestazureblobstoragezerocopyreplication-gw8. Added instance name:node1 tag:8b2301119731 base_cmd:['docker', 'compose', '--env-file', '/ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw8/.env', '--project-name', 'roottestazureblobstoragezerocopyreplication-gw8', '--file', '/ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw8/node1/docker-compose.yml', '--file', '/ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml', '--file', '/ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_azurite.yml'] docker_compose_yml_dir:/ClickHouse/tests/integration/helpers/../../../tests/integration/compose/ clickhouse_start_command: clickhouse server --config-file=/etc/clickhouse-server/{main_config_file} --log-file=/var/log/clickhouse-server/clickhouse-server.log --errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log Cluster name: project_name:roottestazureblobstoragezerocopyreplication-gw8. Added instance name:node2 tag:8b2301119731 base_cmd:['docker', 'compose', '--env-file', '/ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw8/.env', '--project-name', 'roottestazureblobstoragezerocopyreplication-gw8', '--file', '/ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw8/node1/docker-compose.yml', '--file', '/ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml', '--file', '/ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_azurite.yml', '--file', '/ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw8/node2/docker-compose.yml'] docker_compose_yml_dir:/ClickHouse/tests/integration/helpers/../../../tests/integration/compose/ Starting cluster... Running tests in /ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/test.py Cluster start called. is_up=False Docker networks for project roottestazureblobstoragezerocopyreplication-gw8 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestazureblobstoragezerocopyreplication-gw8 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestazureblobstoragezerocopyreplication-gw8 are DRIVER VOLUME NAME Cleanup called Docker networks for project roottestazureblobstoragezerocopyreplication-gw8 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestazureblobstoragezerocopyreplication-gw8 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestazureblobstoragezerocopyreplication-gw8 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestazureblobstoragezerocopyreplication-gw8-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} No running containers for project: roottestazureblobstoragezerocopyreplication-gw8 Trying to prune unused networks... Trying to prune unused images... Command:[docker image prune -f] Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Stdout:5 Command:[docker volume prune -f] Executing query SYSTEM FLUSH LOGS ON CLUSTER 'cluster' on node1 Stdout:Total reclaimed space: 0B Executing query SELECT id, value FROM test_table_backup.test_table on instance Volumes pruned: 5 Setup directory for instance: node1 Create directory for configuration generated in this helper Create directory for common tests configuration Copy common configuration from helpers Generate and write macros file Copy custom test config files ['/ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/configs/config.d/config.xml', '/ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_gen/storage_conf.xml'] to /ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw8/node1/configs/config.d Setup database dir /ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw8/node1/database Setup logs dir /ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw8/node1/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Setup directory for instance: node2 Create directory for configuration generated in this helper Create directory for common tests configuration Copy common configuration from helpers Generate and write macros file Copy custom test config files ['/ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/configs/config.d/config.xml', '/ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_gen/storage_conf.xml'] to /ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw8/node2/configs/config.d Setup database dir /ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw8/node2/database Setup logs dir /ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw8/node2/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw', 'keeper_binary': '/clickhouse', 'keeper_cmd_prefix': 'clickhouse keeper', 'image': 'altinityinfra/integration-test:8b2301119731', 'user': '0', 'keeper_fs': 'bind', 'keeper_logs_dir1': '/ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw8/keeper1/log', 'keeper_config_dir1': '/ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw8/keeper1/config', 'keeper_db_dir1': '/ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw8/keeper1/coordination', 'keeper_logs_dir2': '/ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw8/keeper2/log', 'keeper_config_dir2': '/ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw8/keeper2/config', 'keeper_db_dir2': '/ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw8/keeper2/coordination', 'keeper_logs_dir3': '/ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw8/keeper3/log', 'keeper_config_dir3': '/ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw8/keeper3/config', 'keeper_db_dir3': '/ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw8/keeper3/coordination', 'AZURITE_PORT': '30400', 'AZURITE_STORAGE_ACCOUNT_URL': 'http://azurite1:30400/devstoreaccount1', 'AZURITE_CONNECTION_STRING': 'DefaultEndpointsProtocol=http;AccountName=devstoreaccount1;AccountKey=Eby8vdM02xNOcqFlqUwJPLlmEtlCDXJ1OUzFT50uSRZ6IFsuFq2UVErCz4I6tq/K1SZFPTOtr/KBHBeksoGMGw==;BlobEndpoint=http://azurite1:30400/devstoreaccount1;'} stored in /ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw8/.env Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found http://localhost:None "GET /version HTTP/1.1" 200 826 Command:[docker compose --env-file /ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw8/.env --project-name roottestazureblobstoragezerocopyreplication-gw8 --file /ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw8/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_azurite.yml --file /ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw8/node2/docker-compose.yml pull] Executing query select dictGetInt8('MySQL_complex_key_cache_', 'Int8_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictHas('LocalClickHouse_complex_key_cache_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select dictHas('RedisSimple_direct_Date_', toUInt64(2)) on node Stderr: Container roottestcomposableprotocols-gw9-server-1 Stopping Stderr: Container roottestcomposableprotocols-gw9-server-1 Stopped Stderr: Container roottestcomposableprotocols-gw9-server-1 Removing Stderr: Container roottestcomposableprotocols-gw9-server-1 Removed Stderr: Network roottestcomposableprotocols-gw9_default Removing Stderr: Network roottestcomposableprotocols-gw9_default Removed Cleanup called Docker networks for project roottestcomposableprotocols-gw9 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestcomposableprotocols-gw9 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestcomposableprotocols-gw9 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestcomposableprotocols-gw9-.*-1$' --format '{{.ID}}:{{.Names}}'] run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:True cmd: ['bash', '-c', 'ps -C clickhouse'] Command:[docker exec -u root roottestdatabasebackup-gw1-instance-1 bash -c ps -C clickhouse] Unstopped containers: {} No running containers for project: roottestcomposableprotocols-gw9 Trying to prune unused networks... Executing query select dictHas('MySQL_complex_key_cache_', tuple(toUInt64(2),toString('qwerty2'))) on node Trying to prune unused images... Command:[docker image prune -f] Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Executing query select dictGetOrDefault('LocalClickHouse_complex_key_cache_', 'Float64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat64(777.11)) on local_node Stdout: PID TTY TIME CMD Stdout: 3281 ? 00:00:04 clickhouse run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', 'pkill clickhouse'] Command:[docker exec -u root roottestdatabasebackup-gw1-instance-1 bash -c pkill clickhouse] Stdout:5 Command:[docker volume prune -f] Executing query select dictGetOrDefault('RedisSimple_direct_Date_', 'Date_', toUInt64(9999999), toDate('2018-12-30')) on node Stdout:Total reclaimed space: 0B Volumes pruned: 5 test_config_hide_in_preprocessed/test.py::test_hide_in_preprocessed Running tests in /ClickHouse/tests/integration/test_config_hide_in_preprocessed/test.py Cluster start called. is_up=False run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Docker networks for project roottestconfighideinpreprocessed-gw9 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestconfighideinpreprocessed-gw9 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestconfighideinpreprocessed-gw9 are DRIVER VOLUME NAME Cleanup called Stdout:3281 Docker networks for project roottestconfighideinpreprocessed-gw9 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestconfighideinpreprocessed-gw9 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestconfighideinpreprocessed-gw9 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestconfighideinpreprocessed-gw9-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} No running containers for project: roottestconfighideinpreprocessed-gw9 Trying to prune unused networks... Trying to prune unused images... Command:[docker image prune -f] Executing query select dictGetOrDefault('MySQL_complex_key_cache_', 'Int8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt8(-55)) on node Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Stdout:5 Command:[docker volume prune -f] Stdout:Total reclaimed space: 0B Volumes pruned: 5 Setup directory for instance: node Create directory for configuration generated in this helper Create directory for common tests configuration Copy common configuration from helpers Generate and write macros file Copy custom test config files ['/ClickHouse/tests/integration/test_config_hide_in_preprocessed/configs/config.xml'] to /ClickHouse/tests/integration/test_config_hide_in_preprocessed/_instances-0-gw9/node/configs/config.d Setup database dir /ClickHouse/tests/integration/test_config_hide_in_preprocessed/_instances-0-gw9/node/database Setup logs dir /ClickHouse/tests/integration/test_config_hide_in_preprocessed/_instances-0-gw9/node/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw'} stored in /ClickHouse/tests/integration/test_config_hide_in_preprocessed/_instances-0-gw9/.env Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found http://localhost:None "GET /version HTTP/1.1" 200 826 Command:[docker compose --env-file /ClickHouse/tests/integration/test_config_hide_in_preprocessed/_instances-0-gw9/.env --project-name roottestconfighideinpreprocessed-gw9 --file /ClickHouse/tests/integration/test_config_hide_in_preprocessed/_instances-0-gw9/node/docker-compose.yml pull] Executing query select dictGetFloat64OrDefault('LocalClickHouse_complex_key_cache_', 'Float64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat64(777.11)) on local_node Executing query select dictGetDateOrDefault('RedisSimple_direct_Date_', 'Date_', toUInt64(9999999), toDate('2018-12-30')) on node Executing query select count() from system.tables where database = 'system' and name in ['query_log', 'text_log'] on node2 Executing query select dictGetInt8OrDefault('MySQL_complex_key_cache_', 'Int8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt8(-55)) on node [gw5] PASSED test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_complex[complex_key_cache] test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_complex[complex_key_direct] Executing query system reload dictionaries on local_node Executing query system reload dictionary RedisHash_complex_key_hashed_Date_ on node Executing query rename table system.text_log to system.text_log_20_test on node2 Executing query select dictGet('MySQL_complex_key_cache_', 'Int16_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGet('LocalClickHouse_complex_key_direct_', 'UInt8_', tuple(toUInt64(1),toString('world'))) on local_node Executing query select dictGet('RedisHash_complex_key_hashed_Date_', 'Date_', tuple(toUInt64(1),toString('world'))) on node Executing query rename table system.query_log to system.query_log_20_test on node2 Executing query select dictGetInt16('MySQL_complex_key_cache_', 'Int16_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGetDate('RedisHash_complex_key_hashed_Date_', 'Date_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGetUInt8('LocalClickHouse_complex_key_direct_', 'UInt8_', tuple(toUInt64(1),toString('world'))) on local_node run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] run container_id:roottestasyncloaddatabases-gw7-node2-1 detach:False nothrow:True cmd: ['bash', '-c', 'ps -C clickhouse'] Command:[docker exec -u root roottestasyncloaddatabases-gw7-node2-1 bash -c ps -C clickhouse] Stdout:3281 Stdout: PID TTY TIME CMD Stdout: 787 ? 00:00:25 clickhouse run container_id:roottestasyncloaddatabases-gw7-node2-1 detach:False nothrow:False cmd: ['bash', '-c', 'pkill clickhouse'] Command:[docker exec -u root roottestasyncloaddatabases-gw7-node2-1 bash -c pkill clickhouse] run container_id:roottestasyncloaddatabases-gw7-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw7-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:787 run container_id:roottestbackuprestoreoncluster-gw2-node1-1 detach:False nothrow:False cmd: ['bash', '-c', 'grep -a "Writing backup for file .*kKIdhlkXxt.bin .* (disk default)" /var/log/clickhouse-server/clickhouse-server.log | wc -l'] Command:[docker exec roottestbackuprestoreoncluster-gw2-node1-1 bash -c grep -a "Writing backup for file .*kKIdhlkXxt.bin .* (disk default)" /var/log/clickhouse-server/clickhouse-server.log | wc -l] Executing query select dictHas('MySQL_complex_key_cache_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictHas('RedisHash_complex_key_hashed_Date_', tuple(toUInt64(1),toString('world'))) on node Stdout:1 run container_id:roottestbackuprestoreoncluster-gw2-node2-1 detach:False nothrow:False cmd: ['bash', '-c', 'grep -a "Writing backup for file .*kKIdhlkXxt.bin .* (disk default)" /var/log/clickhouse-server/clickhouse-server.log | wc -l'] Command:[docker exec roottestbackuprestoreoncluster-gw2-node2-1 bash -c grep -a "Writing backup for file .*kKIdhlkXxt.bin .* (disk default)" /var/log/clickhouse-server/clickhouse-server.log | wc -l] Executing query select dictHas('LocalClickHouse_complex_key_direct_', tuple(toUInt64(1),toString('world'))) on local_node Stdout:0 run container_id:roottestbackuprestoreoncluster-gw2-node1-1 detach:False nothrow:False cmd: ['bash', '-c', 'grep -a "Writing backup for file .*kKIdhlkXxt.bin .* skipped" /var/log/clickhouse-server/clickhouse-server.log | wc -l'] Command:[docker exec roottestbackuprestoreoncluster-gw2-node1-1 bash -c grep -a "Writing backup for file .*kKIdhlkXxt.bin .* skipped" /var/log/clickhouse-server/clickhouse-server.log | wc -l] Stdout:3 run container_id:roottestbackuprestoreoncluster-gw2-node2-1 detach:False nothrow:False cmd: ['bash', '-c', 'grep -a "Writing backup for file .*kKIdhlkXxt.bin .* skipped" /var/log/clickhouse-server/clickhouse-server.log | wc -l'] Command:[docker exec roottestbackuprestoreoncluster-gw2-node2-1 bash -c grep -a "Writing backup for file .*kKIdhlkXxt.bin .* skipped" /var/log/clickhouse-server/clickhouse-server.log | wc -l] Stdout:0 Executing query DROP TABLE IF EXISTS tbl ON CLUSTER 'cluster3' SYNC on node1 [gw2] PASSED test_backup_restore_on_cluster/test.py::test_file_deduplication Executing query select dictGetOrDefault('RedisHash_complex_key_hashed_Date_', 'Date_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDate('2018-12-30')) on node Executing query select dictGetOrDefault('MySQL_complex_key_cache_', 'Int16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt16(-66)) on node Executing query select dictGetOrDefault('LocalClickHouse_complex_key_direct_', 'UInt8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt8(55)) on local_node Executing query DROP TABLE IF EXISTS tbl2 ON CLUSTER 'cluster3' SYNC on node1 Executing query select dictGetDateOrDefault('RedisHash_complex_key_hashed_Date_', 'Date_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDate('2018-12-30')) on node Executing query select dictGetInt16OrDefault('MySQL_complex_key_cache_', 'Int16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt16(-66)) on node Executing query select dictGetUInt8OrDefault('LocalClickHouse_complex_key_direct_', 'UInt8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt8(55)) on local_node run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select dictGet('MySQL_complex_key_cache_', 'Int32_', tuple(toUInt64(2),toString('qwerty2'))) on node Stdout:3281 Executing query select dictGet('RedisHash_complex_key_hashed_Date_', 'Date_', tuple(toUInt64(2),toString('qwerty2'))) on node run container_id:roottestasyncloaddatabases-gw7-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw7-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query DROP DATABASE IF EXISTS mydb ON CLUSTER 'cluster3' SYNC on node1 Stdout:787 Executing query select dictGet('LocalClickHouse_complex_key_direct_', 'UInt16_', tuple(toUInt64(1),toString('world'))) on local_node Executing query select dictGetInt32('MySQL_complex_key_cache_', 'Int32_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGetDate('RedisHash_complex_key_hashed_Date_', 'Date_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGetUInt16('LocalClickHouse_complex_key_direct_', 'UInt16_', tuple(toUInt64(1),toString('world'))) on local_node Executing query DROP DATABASE IF EXISTS mydb2 ON CLUSTER 'cluster3' SYNC on node1 Executing query select dictHas('MySQL_complex_key_cache_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictHas('RedisHash_complex_key_hashed_Date_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictHas('LocalClickHouse_complex_key_direct_', tuple(toUInt64(1),toString('world'))) on local_node Executing query select dictGetOrDefault('MySQL_complex_key_cache_', 'Int32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt32(-77)) on node Executing query select dictGetOrDefault('RedisHash_complex_key_hashed_Date_', 'Date_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDate('2018-12-30')) on node Executing query DROP USER IF EXISTS u1, u2 ON CLUSTER 'cluster3' on node1 run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select dictGetInt32OrDefault('MySQL_complex_key_cache_', 'Int32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt32(-77)) on node Stdout:3281 Executing query select dictGetOrDefault('LocalClickHouse_complex_key_direct_', 'UInt16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt16(66)) on local_node Executing query select dictGetDateOrDefault('RedisHash_complex_key_hashed_Date_', 'Date_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDate('2018-12-30')) on node run container_id:roottestasyncloaddatabases-gw7-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw7-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:787 test_backup_restore_on_cluster/test.py::test_get_error_from_other_host Executing query CREATE TABLE tbl (`x` UInt8) ENGINE = MergeTree ORDER BY x on node1 Executing query system reload dictionary RedisHash_complex_key_cache_Date_ on node Executing query select dictGet('MySQL_complex_key_cache_', 'Int64_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGetUInt16OrDefault('LocalClickHouse_complex_key_direct_', 'UInt16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt16(66)) on local_node Executing query INSERT INTO tbl VALUES (3) on node1 Executing query select dictGet('RedisHash_complex_key_cache_Date_', 'Date_', tuple(toUInt64(1),toString('world'))) on node Executing query BACKUP TABLE tbl ON CLUSTER 'cluster' TO Disk('backups', '8') on node1 Executing query select dictGetInt64('MySQL_complex_key_cache_', 'Int64_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGet('LocalClickHouse_complex_key_direct_', 'UInt32_', tuple(toUInt64(1),toString('world'))) on local_node Executing query select dictGetDate('RedisHash_complex_key_cache_Date_', 'Date_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictHas('MySQL_complex_key_cache_', tuple(toUInt64(2),toString('qwerty2'))) on node run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select dictGetUInt32('LocalClickHouse_complex_key_direct_', 'UInt32_', tuple(toUInt64(1),toString('world'))) on local_node run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select dictHas('RedisHash_complex_key_cache_Date_', tuple(toUInt64(1),toString('world'))) on node No clickhouse process running. Start new one. run container_id:roottestasyncloaddatabases-gw7-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] http://localhost:None "POST /v1.46/containers/roottestdatabasebackup-gw1-instance-1/exec HTTP/1.1" 201 74 Command:[docker exec roottestasyncloaddatabases-gw7-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] http://localhost:None "POST /v1.46/exec/bb3b7bd4912f32c87b7eb4539ae1d27a9217d3403e434a55cad70e5e9df34f70/start HTTP/1.1" 200 0 http://localhost:None "GET /v1.46/exec/bb3b7bd4912f32c87b7eb4539ae1d27a9217d3403e434a55cad70e5e9df34f70/json HTTP/1.1" 200 586 Stdout:787 Executing query select dictGetOrDefault('MySQL_complex_key_cache_', 'Int64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt64(-88)) on node Executing query select dictHas('LocalClickHouse_complex_key_direct_', tuple(toUInt64(1),toString('world'))) on local_node Executing query DROP TABLE IF EXISTS tbl ON CLUSTER 'cluster3' SYNC on node1 [gw2] PASSED test_backup_restore_on_cluster/test.py::test_get_error_from_other_host Executing query select dictGetOrDefault('RedisHash_complex_key_cache_Date_', 'Date_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDate('2018-12-30')) on node Executing query select dictGetInt64OrDefault('MySQL_complex_key_cache_', 'Int64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt64(-88)) on node Executing query select dictGetOrDefault('LocalClickHouse_complex_key_direct_', 'UInt32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt32(77)) on local_node Executing query select dictGetDateOrDefault('RedisHash_complex_key_cache_Date_', 'Date_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDate('2018-12-30')) on node Executing query select dictGet('MySQL_complex_key_cache_', 'UUID_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query DROP TABLE IF EXISTS tbl2 ON CLUSTER 'cluster3' SYNC on node1 Executing query select dictGetUInt32OrDefault('LocalClickHouse_complex_key_direct_', 'UInt32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt32(77)) on local_node Executing query select dictGet('RedisHash_complex_key_cache_Date_', 'Date_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGetUUID('MySQL_complex_key_cache_', 'UUID_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query DROP DATABASE IF EXISTS mydb ON CLUSTER 'cluster3' SYNC on node1 run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] run container_id:roottestasyncloaddatabases-gw7-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw7-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:4111 Clickhouse process running. run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] run container_id:roottestasyncloaddatabases-gw7-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw7-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:4111 Executing query select 20 on instance No clickhouse process running. Start new one. http://localhost:None "POST /v1.46/containers/roottestasyncloaddatabases-gw7-node2-1/exec HTTP/1.1" 201 74 http://localhost:None "POST /v1.46/exec/2c37df98ba4ecff02ab033601a6f69a1ed2419d9267e2feb9aaaf817cccda99a/start HTTP/1.1" 200 0 http://localhost:None "GET /v1.46/exec/2c37df98ba4ecff02ab033601a6f69a1ed2419d9267e2feb9aaaf817cccda99a/json HTTP/1.1" 200 586 Executing query select dictGetDate('RedisHash_complex_key_cache_Date_', 'Date_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictHas('MySQL_complex_key_cache_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGet('LocalClickHouse_complex_key_direct_', 'UInt64_', tuple(toUInt64(1),toString('world'))) on local_node Executing query DROP DATABASE IF EXISTS mydb2 ON CLUSTER 'cluster3' SYNC on node1 Executing query select dictHas('RedisHash_complex_key_cache_Date_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGetOrDefault('MySQL_complex_key_cache_', 'UUID_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUUID('550e8400-0000-0000-0000-000000000000')) on node Executing query select dictGetUInt64('LocalClickHouse_complex_key_direct_', 'UInt64_', tuple(toUInt64(1),toString('world'))) on local_node Executing query select dictGetOrDefault('RedisHash_complex_key_cache_Date_', 'Date_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDate('2018-12-30')) on node Executing query DROP USER IF EXISTS u1, u2 ON CLUSTER 'cluster3' on node1 Executing query select dictGetUUIDOrDefault('MySQL_complex_key_cache_', 'UUID_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUUID('550e8400-0000-0000-0000-000000000000')) on node Executing query select 20 on instance Executing query select dictHas('LocalClickHouse_complex_key_direct_', tuple(toUInt64(1),toString('world'))) on local_node Executing query select dictGetDateOrDefault('RedisHash_complex_key_cache_Date_', 'Date_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDate('2018-12-30')) on node Executing query select dictGet('MySQL_complex_key_cache_', 'Date_', tuple(toUInt64(2),toString('qwerty2'))) on node test_backup_restore_on_cluster/test.py::test_keeper_value_max_size Executing query CREATE TABLE tbl ON CLUSTER 'cluster' (x UInt32) ENGINE=ReplicatedMergeTree('/clickhouse/tables/tbl/', '{replica}')ORDER BY x on node1 run container_id:roottestasyncloaddatabases-gw7-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw7-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select dictGetOrDefault('LocalClickHouse_complex_key_direct_', 'UInt64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt64(88)) on local_node Stdout:1607 Clickhouse process running. run container_id:roottestasyncloaddatabases-gw7-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw7-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:1607 Executing query select 20 on node2 Executing query select dictGetDate('MySQL_complex_key_cache_', 'Date_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query system reload dictionary RedisHash_complex_key_direct_Date_ on node Executing query select 20 on instance Executing query select dictGetUInt64OrDefault('LocalClickHouse_complex_key_direct_', 'UInt64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt64(88)) on local_node Executing query INSERT INTO tbl VALUES (111) on node1 Executing query select dictHas('MySQL_complex_key_cache_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGet('RedisHash_complex_key_direct_Date_', 'Date_', tuple(toUInt64(1),toString('world'))) on node Executing query INSERT INTO tbl VALUES (222) on node2 Executing query select dictGet('LocalClickHouse_complex_key_direct_', 'Int8_', tuple(toUInt64(1),toString('world'))) on local_node Executing query select dictGetOrDefault('MySQL_complex_key_cache_', 'Date_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDate('2018-12-30')) on node Executing query select dictGetDate('RedisHash_complex_key_direct_Date_', 'Date_', tuple(toUInt64(1),toString('world'))) on node Executing query select 20 on node2 Executing query SYSTEM SYNC REPLICA ON CLUSTER 'cluster' tbl on node1 Executing query select dictGetDateOrDefault('MySQL_complex_key_cache_', 'Date_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDate('2018-12-30')) on node Executing query select dictHas('RedisHash_complex_key_direct_Date_', tuple(toUInt64(1),toString('world'))) on node Stderr: node Pulling Stderr: node Pulled ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_attach_table_normalizer/_instances-0-gw6/.env --project-name roottestattachtablenormalizer-gw6 --file /ClickHouse/tests/integration/test_attach_table_normalizer/_instances-0-gw6/node/docker-compose.yml up -d --no-recreate') Command:[docker compose --env-file /ClickHouse/tests/integration/test_attach_table_normalizer/_instances-0-gw6/.env --project-name roottestattachtablenormalizer-gw6 --file /ClickHouse/tests/integration/test_attach_table_normalizer/_instances-0-gw6/node/docker-compose.yml up -d --no-recreate] Executing query select dictGetInt8('LocalClickHouse_complex_key_direct_', 'Int8_', tuple(toUInt64(1),toString('world'))) on local_node Executing query select 20 on instance Stderr: node Pulling Stderr: node Pulled ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_config_hide_in_preprocessed/_instances-0-gw9/.env --project-name roottestconfighideinpreprocessed-gw9 --file /ClickHouse/tests/integration/test_config_hide_in_preprocessed/_instances-0-gw9/node/docker-compose.yml up -d --no-recreate') Command:[docker compose --env-file /ClickHouse/tests/integration/test_config_hide_in_preprocessed/_instances-0-gw9/.env --project-name roottestconfighideinpreprocessed-gw9 --file /ClickHouse/tests/integration/test_config_hide_in_preprocessed/_instances-0-gw9/node/docker-compose.yml up -d --no-recreate] Executing query select dictGet('MySQL_complex_key_cache_', 'DateTime_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGetOrDefault('RedisHash_complex_key_direct_Date_', 'Date_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDate('2018-12-30')) on node Executing query SYSTEM STOP REPLICATED SENDS ON CLUSTER 'cluster' tbl on node1 Executing query select dictHas('LocalClickHouse_complex_key_direct_', tuple(toUInt64(1),toString('world'))) on local_node Executing query SELECT id, value FROM test_table_backup.test_table on instance Executing query select 20 on node2 Executing query select dictGetDateTime('MySQL_complex_key_cache_', 'DateTime_', tuple(toUInt64(2),toString('qwerty2'))) on node Stderr: Network roottestattachtablenormalizer-gw6_default Creating Stderr: Network roottestattachtablenormalizer-gw6_default Created Stderr: Container roottestattachtablenormalizer-gw6-node-1 Creating Stderr: Container roottestattachtablenormalizer-gw6-node-1 Created Stderr: Container roottestattachtablenormalizer-gw6-node-1 Starting Stderr: Container roottestattachtablenormalizer-gw6-node-1 Started ClickHouse instance created get_instance_ip instance_name=node http://localhost:None "GET /v1.46/containers/roottestattachtablenormalizer-gw6-node-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node Executing query select dictGetDateOrDefault('RedisHash_complex_key_direct_Date_', 'Date_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDate('2018-12-30')) on node http://localhost:None "GET /v1.46/containers/roottestattachtablenormalizer-gw6-node-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node, ip: 172.16.1.2... http://localhost:None "GET /v1.46/containers/roottestattachtablenormalizer-gw6-node-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/b85c6a214b6eed5230a0bbb732277e57d8d74ddb5e1ff32466d45e4c84bf06af/json HTTP/1.1" 200 None Executing query select dictGetOrDefault('LocalClickHouse_complex_key_direct_', 'Int8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt8(-55)) on local_node Executing query INSERT INTO tbl VALUES (333) on node1 http://localhost:None "GET /v1.46/containers/b85c6a214b6eed5230a0bbb732277e57d8d74ddb5e1ff32466d45e4c84bf06af/json HTTP/1.1" 200 None Executing query DROP DATABASE test_table_backup on instance http://localhost:None "GET /v1.46/containers/b85c6a214b6eed5230a0bbb732277e57d8d74ddb5e1ff32466d45e4c84bf06af/json HTTP/1.1" 200 None Executing query select dictGet('RedisHash_complex_key_direct_Date_', 'Date_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictHas('MySQL_complex_key_cache_', tuple(toUInt64(2),toString('qwerty2'))) on node Stderr: Network roottestconfighideinpreprocessed-gw9_default Creating Stderr: Network roottestconfighideinpreprocessed-gw9_default Created Stderr: Container roottestconfighideinpreprocessed-gw9-node-1 Creating Stderr: Container roottestconfighideinpreprocessed-gw9-node-1 Created Stderr: Container roottestconfighideinpreprocessed-gw9-node-1 Starting Stderr: Container roottestconfighideinpreprocessed-gw9-node-1 Started ClickHouse instance created get_instance_ip instance_name=node http://localhost:None "GET /v1.46/containers/roottestconfighideinpreprocessed-gw9-node-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node http://localhost:None "GET /v1.46/containers/roottestconfighideinpreprocessed-gw9-node-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node, ip: 172.16.4.2... http://localhost:None "GET /v1.46/containers/roottestconfighideinpreprocessed-gw9-node-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/9c05e683d3e957fc1337371f5823e44dbf36663c096918655bdcba069f3c3cc1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/b85c6a214b6eed5230a0bbb732277e57d8d74ddb5e1ff32466d45e4c84bf06af/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/9c05e683d3e957fc1337371f5823e44dbf36663c096918655bdcba069f3c3cc1/json HTTP/1.1" 200 None Executing query DROP DATABASE test_database on instance http://localhost:None "GET /v1.46/containers/b85c6a214b6eed5230a0bbb732277e57d8d74ddb5e1ff32466d45e4c84bf06af/json HTTP/1.1" 200 None Executing query INSERT INTO tbl VALUES (444) on node2 Executing query select dictGetInt8OrDefault('LocalClickHouse_complex_key_direct_', 'Int8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt8(-55)) on local_node http://localhost:None "GET /v1.46/containers/9c05e683d3e957fc1337371f5823e44dbf36663c096918655bdcba069f3c3cc1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/b85c6a214b6eed5230a0bbb732277e57d8d74ddb5e1ff32466d45e4c84bf06af/json HTTP/1.1" 200 None Executing query select dictGetDate('RedisHash_complex_key_direct_Date_', 'Date_', tuple(toUInt64(2),toString('qwerty2'))) on node http://localhost:None "GET /v1.46/containers/9c05e683d3e957fc1337371f5823e44dbf36663c096918655bdcba069f3c3cc1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/b85c6a214b6eed5230a0bbb732277e57d8d74ddb5e1ff32466d45e4c84bf06af/json HTTP/1.1" 200 None run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', 'rm -rf /backups/'] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c rm -rf /backups/] Executing query select 20 on node2 Executing query select dictGetOrDefault('MySQL_complex_key_cache_', 'DateTime_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDateTime('2018-12-30 00:00:00')) on node Executing query BACKUP TABLE tbl ON CLUSTER 'cluster' TO Disk('backups', '9') on node1 http://localhost:None "GET /v1.46/containers/9c05e683d3e957fc1337371f5823e44dbf36663c096918655bdcba069f3c3cc1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/b85c6a214b6eed5230a0bbb732277e57d8d74ddb5e1ff32466d45e4c84bf06af/json HTTP/1.1" 200 None run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', 'rm -rf /local_plain/'] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c rm -rf /local_plain/] http://172.16.8.4:9001 "GET /root?delimiter=&encoding-type=url&list-type=2&max-keys=1000&prefix= HTTP/1.1" 200 0 run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', 'rm -rf /backups/'] [gw1] PASSED test_database_backup/test.py::test_database_backup_table[Disk('backup_disk_local', 'test_table_backup')] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c rm -rf /backups/] test_database_backup/test.py::test_database_backup_table[Disk('backup_disk_object_storage_local_plain', 'test_table_backup')] http://localhost:None "GET /v1.46/containers/9c05e683d3e957fc1337371f5823e44dbf36663c096918655bdcba069f3c3cc1/json HTTP/1.1" 200 None Executing query select dictGet('LocalClickHouse_complex_key_direct_', 'Int16_', tuple(toUInt64(1),toString('world'))) on local_node http://localhost:None "GET /v1.46/containers/b85c6a214b6eed5230a0bbb732277e57d8d74ddb5e1ff32466d45e4c84bf06af/json HTTP/1.1" 200 None run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', 'rm -rf /local_plain/'] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c rm -rf /local_plain/] http://localhost:None "GET /v1.46/containers/9c05e683d3e957fc1337371f5823e44dbf36663c096918655bdcba069f3c3cc1/json HTTP/1.1" 200 None Executing query select dictHas('RedisHash_complex_key_direct_Date_', tuple(toUInt64(2),toString('qwerty2'))) on node http://localhost:None "GET /v1.46/containers/b85c6a214b6eed5230a0bbb732277e57d8d74ddb5e1ff32466d45e4c84bf06af/json HTTP/1.1" 200 None http://172.16.8.4:9001 "GET /root?delimiter=&encoding-type=url&list-type=2&max-keys=1000&prefix= HTTP/1.1" 200 0 Executing query DROP DATABASE IF EXISTS test_database; DROP DATABASE IF EXISTS test_table_backup; CREATE DATABASE test_database; CREATE TABLE test_database.test_table (id UInt64, value String) ENGINE=MergeTree ORDER BY id; INSERT INTO test_database.test_table VALUES (0, 'test_database.test_table'); BACKUP TABLE test_database.test_table TO Disk('backup_disk_object_storage_local_plain', 'test_table_backup'); CREATE DATABASE test_table_backup ENGINE = Backup('test_database', Disk('backup_disk_object_storage_local_plain', 'test_table_backup')); on instance Executing query select dictGetDateTimeOrDefault('MySQL_complex_key_cache_', 'DateTime_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDateTime('2018-12-30 00:00:00')) on node http://localhost:None "GET /v1.46/containers/9c05e683d3e957fc1337371f5823e44dbf36663c096918655bdcba069f3c3cc1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/b85c6a214b6eed5230a0bbb732277e57d8d74ddb5e1ff32466d45e4c84bf06af/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/9c05e683d3e957fc1337371f5823e44dbf36663c096918655bdcba069f3c3cc1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/b85c6a214b6eed5230a0bbb732277e57d8d74ddb5e1ff32466d45e4c84bf06af/json HTTP/1.1" 200 None Executing query select dictGetInt16('LocalClickHouse_complex_key_direct_', 'Int16_', tuple(toUInt64(1),toString('world'))) on local_node Executing query select dictGetOrDefault('RedisHash_complex_key_direct_Date_', 'Date_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDate('2018-12-30')) on node http://localhost:None "GET /v1.46/containers/9c05e683d3e957fc1337371f5823e44dbf36663c096918655bdcba069f3c3cc1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/b85c6a214b6eed5230a0bbb732277e57d8d74ddb5e1ff32466d45e4c84bf06af/json HTTP/1.1" 200 None Executing query select dictGet('MySQL_complex_key_cache_', 'String_', tuple(toUInt64(2),toString('qwerty2'))) on node http://localhost:None "GET /v1.46/containers/9c05e683d3e957fc1337371f5823e44dbf36663c096918655bdcba069f3c3cc1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/b85c6a214b6eed5230a0bbb732277e57d8d74ddb5e1ff32466d45e4c84bf06af/json HTTP/1.1" 200 None Executing query select 20 on node2 Command:[docker compose --env-file /ClickHouse/tests/integration/test_cluster_all_replicas/_instances-0-gw4/.env --project-name roottestclusterallreplicas-gw4 --file /ClickHouse/tests/integration/test_cluster_all_replicas/_instances-0-gw4/node1/docker-compose.yml --file /ClickHouse/tests/integration/test_cluster_all_replicas/_instances-0-gw4/node2/docker-compose.yml stop --timeout 20] [gw4] PASSED test_cluster_all_replicas/test.py::test_skip_unavailable_replica[two_shards_three_nodes] Executing query SELECT id, value FROM test_table_backup.test_table on instance http://localhost:None "GET /v1.46/containers/9c05e683d3e957fc1337371f5823e44dbf36663c096918655bdcba069f3c3cc1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/b85c6a214b6eed5230a0bbb732277e57d8d74ddb5e1ff32466d45e4c84bf06af/json HTTP/1.1" 200 None Executing query select dictHas('LocalClickHouse_complex_key_direct_', tuple(toUInt64(1),toString('world'))) on local_node Executing query select dictGetDateOrDefault('RedisHash_complex_key_direct_Date_', 'Date_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDate('2018-12-30')) on node http://localhost:None "GET /v1.46/containers/9c05e683d3e957fc1337371f5823e44dbf36663c096918655bdcba069f3c3cc1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/b85c6a214b6eed5230a0bbb732277e57d8d74ddb5e1ff32466d45e4c84bf06af/json HTTP/1.1" 200 None Executing query select dictGetString('MySQL_complex_key_cache_', 'String_', tuple(toUInt64(2),toString('qwerty2'))) on node http://localhost:None "GET /v1.46/containers/9c05e683d3e957fc1337371f5823e44dbf36663c096918655bdcba069f3c3cc1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/b85c6a214b6eed5230a0bbb732277e57d8d74ddb5e1ff32466d45e4c84bf06af/json HTTP/1.1" 200 None run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:True cmd: ['bash', '-c', 'ps -C clickhouse'] Command:[docker exec -u root roottestdatabasebackup-gw1-instance-1 bash -c ps -C clickhouse] http://localhost:None "GET /v1.46/containers/9c05e683d3e957fc1337371f5823e44dbf36663c096918655bdcba069f3c3cc1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/b85c6a214b6eed5230a0bbb732277e57d8d74ddb5e1ff32466d45e4c84bf06af/json HTTP/1.1" 200 None Stdout: PID TTY TIME CMD Stdout: 4111 ? 00:00:04 clickhouse run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', 'pkill clickhouse'] Command:[docker exec -u root roottestdatabasebackup-gw1-instance-1 bash -c pkill clickhouse] Executing query select dictGetOrDefault('LocalClickHouse_complex_key_direct_', 'Int16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt16(-66)) on local_node Executing query system reload dictionaries on node run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] http://localhost:None "GET /v1.46/containers/9c05e683d3e957fc1337371f5823e44dbf36663c096918655bdcba069f3c3cc1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/b85c6a214b6eed5230a0bbb732277e57d8d74ddb5e1ff32466d45e4c84bf06af/json HTTP/1.1" 200 None Stdout:4111 Executing query select dictHas('MySQL_complex_key_cache_', tuple(toUInt64(2),toString('qwerty2'))) on node http://localhost:None "GET /v1.46/containers/9c05e683d3e957fc1337371f5823e44dbf36663c096918655bdcba069f3c3cc1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/b85c6a214b6eed5230a0bbb732277e57d8d74ddb5e1ff32466d45e4c84bf06af/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/9c05e683d3e957fc1337371f5823e44dbf36663c096918655bdcba069f3c3cc1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/b85c6a214b6eed5230a0bbb732277e57d8d74ddb5e1ff32466d45e4c84bf06af/json HTTP/1.1" 200 None Executing query select dictGetInt16OrDefault('LocalClickHouse_complex_key_direct_', 'Int16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt16(-66)) on local_node Run test with id: 12 Executing query system reload dictionaries on node [gw0] PASSED test_dictionaries_redis/test.py::test_redis_dictionaries[Date_] test_dictionaries_redis/test.py::test_redis_dictionaries[Float32_] http://localhost:None "GET /v1.46/containers/9c05e683d3e957fc1337371f5823e44dbf36663c096918655bdcba069f3c3cc1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/b85c6a214b6eed5230a0bbb732277e57d8d74ddb5e1ff32466d45e4c84bf06af/json HTTP/1.1" 200 None Executing query select dictGetOrDefault('MySQL_complex_key_cache_', 'String_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toString('hi')) on node http://localhost:None "GET /v1.46/containers/9c05e683d3e957fc1337371f5823e44dbf36663c096918655bdcba069f3c3cc1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/b85c6a214b6eed5230a0bbb732277e57d8d74ddb5e1ff32466d45e4c84bf06af/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/9c05e683d3e957fc1337371f5823e44dbf36663c096918655bdcba069f3c3cc1/json HTTP/1.1" 200 None Executing query system reload dictionary RedisSimple_flat_Float32_ on node Executing query DROP TABLE tbl ON CLUSTER 'cluster' SYNC on node1 http://localhost:None "GET /v1.46/containers/b85c6a214b6eed5230a0bbb732277e57d8d74ddb5e1ff32466d45e4c84bf06af/json HTTP/1.1" 200 None Executing query select dictGet('LocalClickHouse_complex_key_direct_', 'Int32_', tuple(toUInt64(1),toString('world'))) on local_node http://localhost:None "GET /v1.46/containers/9c05e683d3e957fc1337371f5823e44dbf36663c096918655bdcba069f3c3cc1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/b85c6a214b6eed5230a0bbb732277e57d8d74ddb5e1ff32466d45e4c84bf06af/json HTTP/1.1" 200 None Executing query select dictGetStringOrDefault('MySQL_complex_key_cache_', 'String_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toString('hi')) on node http://localhost:None "GET /v1.46/containers/9c05e683d3e957fc1337371f5823e44dbf36663c096918655bdcba069f3c3cc1/json HTTP/1.1" 200 None Executing query select dictGet('RedisSimple_flat_Float32_', 'Float32_', toUInt64(1)) on node http://localhost:None "GET /v1.46/containers/b85c6a214b6eed5230a0bbb732277e57d8d74ddb5e1ff32466d45e4c84bf06af/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/9c05e683d3e957fc1337371f5823e44dbf36663c096918655bdcba069f3c3cc1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/b85c6a214b6eed5230a0bbb732277e57d8d74ddb5e1ff32466d45e4c84bf06af/json HTTP/1.1" 200 None Executing query select dictGetInt32('LocalClickHouse_complex_key_direct_', 'Int32_', tuple(toUInt64(1),toString('world'))) on local_node http://localhost:None "GET /v1.46/containers/9c05e683d3e957fc1337371f5823e44dbf36663c096918655bdcba069f3c3cc1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/b85c6a214b6eed5230a0bbb732277e57d8d74ddb5e1ff32466d45e4c84bf06af/json HTTP/1.1" 200 None Executing query select dictGetFloat32('RedisSimple_flat_Float32_', 'Float32_', toUInt64(1)) on node http://localhost:None "GET /v1.46/containers/9c05e683d3e957fc1337371f5823e44dbf36663c096918655bdcba069f3c3cc1/json HTTP/1.1" 200 None Executing query select dictGet('MySQL_complex_key_cache_', 'Float32_', tuple(toUInt64(2),toString('qwerty2'))) on node run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] http://localhost:None "GET /v1.46/containers/b85c6a214b6eed5230a0bbb732277e57d8d74ddb5e1ff32466d45e4c84bf06af/json HTTP/1.1" 200 None Stdout:4111 http://localhost:None "GET /v1.46/containers/9c05e683d3e957fc1337371f5823e44dbf36663c096918655bdcba069f3c3cc1/json HTTP/1.1" 200 None Executing query RESTORE TABLE tbl ON CLUSTER 'cluster' FROM Disk('backups', '9') on node1 http://localhost:None "GET /v1.46/containers/b85c6a214b6eed5230a0bbb732277e57d8d74ddb5e1ff32466d45e4c84bf06af/json HTTP/1.1" 200 None ClickHouse node started Executing query DROP TABLE IF EXISTS default.file on node Executing query select dictHas('LocalClickHouse_complex_key_direct_', tuple(toUInt64(1),toString('world'))) on local_node http://localhost:None "GET /v1.46/containers/9c05e683d3e957fc1337371f5823e44dbf36663c096918655bdcba069f3c3cc1/json HTTP/1.1" 200 None ClickHouse node started Executing query select value from system.server_settings where name ='max_thread_pool_free_size' on node Executing query select dictHas('RedisSimple_flat_Float32_', toUInt64(1)) on node Executing query select dictGetFloat32('MySQL_complex_key_cache_', 'Float32_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query CREATE TABLE default.file(`s` String, `n` UInt8) ENGINE = MergeTree PARTITION BY substring(s, 1, 2) ORDER BY n on node Executing query select dictGetOrDefault('LocalClickHouse_complex_key_direct_', 'Int32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt32(-77)) on local_node Executing query select dictGetOrDefault('RedisSimple_flat_Float32_', 'Float32_', toUInt64(9999999), toFloat32(555.11)) on node Executing query DETACH TABLE file on node Executing query select value from system.server_settings where name ='max_table_size_to_drop' on node Executing query select dictHas('MySQL_complex_key_cache_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select count() from system.text_log_12_test on node2 run container_id:roottestattachtablenormalizer-gw6-node-1 detach:False nothrow:False cmd: ['bash', '-c', "sed -i 's/substring/substr/g' /var/lib/clickhouse/metadata/default/file.sql"] Command:[docker exec -u root roottestattachtablenormalizer-gw6-node-1 bash -c sed -i 's/substring/substr/g' /var/lib/clickhouse/metadata/default/file.sql] Executing query select dictGetFloat32OrDefault('RedisSimple_flat_Float32_', 'Float32_', toUInt64(9999999), toFloat32(555.11)) on node Executing query select dictGetInt32OrDefault('LocalClickHouse_complex_key_direct_', 'Int32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt32(-77)) on local_node Executing query ATTACH TABLE file on node Executing query select value from system.server_settings where name ='max_partition_size_to_drop' on node Executing query select dictGetOrDefault('MySQL_complex_key_cache_', 'Float32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat32(555.11)) on node Executing query select count() from system.query_log_8_test on node2 [gw6] PASSED test_attach_table_normalizer/test.py::test_attach_substr Executing query DROP TABLE IF EXISTS default.file on node test_attach_table_normalizer/test.py::test_attach_substr_restart Executing query select dictGet('RedisSimple_flat_Float32_', 'Float32_', toUInt64(2)) on node run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select collection from system.named_collections on node Executing query select dictGet('LocalClickHouse_complex_key_direct_', 'Int64_', tuple(toUInt64(1),toString('world'))) on local_node Executing query select dictGetFloat32OrDefault('MySQL_complex_key_cache_', 'Float32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat32(555.11)) on node Stdout:4111 Executing query select count() from system.text_log_11_test on node2 Executing query CREATE TABLE default.file(`s` String, `n` UInt8) ENGINE = MergeTree PARTITION BY substring(s, 1, 2) ORDER BY n on node Executing query select dictGetFloat32('RedisSimple_flat_Float32_', 'Float32_', toUInt64(2)) on node Executing query select dictGet('MySQL_complex_key_cache_', 'Float64_', tuple(toUInt64(2),toString('qwerty2'))) on node run container_id:roottestconfighideinpreprocessed-gw9-node-1 detach:False nothrow:False cmd: ['cat', '/var/lib/clickhouse/preprocessed_configs/config.xml'] Command:[docker exec roottestconfighideinpreprocessed-gw9-node-1 cat /var/lib/clickhouse/preprocessed_configs/config.xml] Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: test Stdout: /var/log/clickhouse-server/clickhouse-server.log Stdout: /var/log/clickhouse-server/clickhouse-server.err.log Stdout: Stdout: 1000M Stdout: 10 Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: /var/log/clickhouse-server/stderr.log Stdout: /var/log/clickhouse-server/stdout.log Stdout: true Stdout: Stdout: Stdout: Stdout: Stdout: https://{bucket}.s3.amazonaws.com Stdout: Stdout: Stdout: https://storage.googleapis.com/{bucket} Stdout: Stdout: Stdout: https://{bucket}.oss.aliyuncs.com Stdout: Stdout: Stdout: Stdout: Stdout: Stdout:
Stdout: Access-Control-Allow-Origin Stdout: * Stdout:
Stdout:
Stdout: Access-Control-Allow-Headers Stdout: origin, x-requested-with, x-clickhouse-format, x-clickhouse-user, x-clickhouse-key, Authorization Stdout:
Stdout:
Stdout: Access-Control-Allow-Methods Stdout: POST, GET, OPTIONS Stdout:
Stdout:
Stdout: Access-Control-Max-Age Stdout: 86400 Stdout:
Stdout:
Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: 8123 Stdout: Stdout: Stdout: 9000 Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: 9004 Stdout: Stdout: Stdout: 9005 Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: 9009 Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Executing query SYSTEM SYNC REPLICA ON CLUSTER 'cluster' tbl on node1 Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: 10 Stdout: Stdout: Stdout: Stdout: Stdout: false Stdout: Stdout: Stdout: /path/to/ssl_cert_file Stdout: /path/to/ssl_key_file Stdout: Stdout: Stdout: false Stdout: Stdout: Stdout: /path/to/ssl_ca_cert_file Stdout: Stdout: Stdout: none Stdout: Stdout: Stdout: 0 Stdout: Stdout: Stdout: -1 Stdout: -1 Stdout: Stdout: Stdout: false Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: none Stdout: true Stdout: true Stdout: sslv2,sslv3 Stdout: true Stdout: Stdout: Stdout: Stdout: RejectCertificateHandler Stdout: Stdout: Stdout: Stdout: Stdout: true Stdout: true Stdout: sslv2,sslv3 Stdout: true Stdout: Stdout: Stdout: Stdout: RejectCertificateHandler Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: 0 Stdout: 2 Stdout: fair_round_robin Stdout: Stdout: Stdout: 1000 Stdout: Stdout: Stdout: 0 Stdout: Stdout: Stdout: Stdout: 10000 Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: true Stdout: Stdout: Stdout: 0.9 Stdout: Stdout: Stdout: 4194304 Stdout: Stdout: Stdout: 0 Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: 8589934592 Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: / Stdout: Stdout: false Stdout: Stdout: Stdout: /var/lib/clickhouse/ Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: /var/lib/clickhouse/tmp/ Stdout: Stdout: Stdout: 1 Stdout: 1 Stdout: 1 Stdout: Stdout: Stdout: sha256_password Stdout: Stdout: Stdout: 12 Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: /var/lib/clickhouse/user_files/ Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: users.xml Stdout: Stdout: Stdout: Stdout: /var/lib/clickhouse/access/ Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: true Stdout: Stdout: Stdout: true Stdout: Stdout: Stdout: true Stdout: Stdout: Stdout: true Stdout: Stdout: Stdout: true Stdout: Stdout: Stdout: false Stdout: Stdout: Stdout: 2 Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: default Stdout: Stdout: Stdout: custom_ Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: default Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: true Stdout: Stdout: Stdout: false Stdout: Stdout: ' | sed -e 's|.*>\(.*\)<.*|\1|') Stdout: wget https://github.com/ClickHouse/clickhouse-jdbc-bridge/releases/download/v$PKG_VER/clickhouse-jdbc-bridge_$PKG_VER-1_all.deb Stdout: apt install --no-install-recommends -f ./clickhouse-jdbc-bridge_$PKG_VER-1_all.deb Stdout: clickhouse-jdbc-bridge & Stdout: Stdout: * [CentOS/RHEL] Stdout: export MVN_URL=https://repo1.maven.org/maven2/com/clickhouse/clickhouse-jdbc-bridge/ Stdout: export PKG_VER=$(curl -sL $MVN_URL/maven-metadata.xml | grep '' | sed -e 's|.*>\(.*\)<.*|\1|') Stdout: wget https://github.com/ClickHouse/clickhouse-jdbc-bridge/releases/download/v$PKG_VER/clickhouse-jdbc-bridge-$PKG_VER-1.noarch.rpm Stdout: yum localinstall -y clickhouse-jdbc-bridge-$PKG_VER-1.noarch.rpm Stdout: clickhouse-jdbc-bridge & Stdout: Stdout: Please refer to https://github.com/ClickHouse/clickhouse-jdbc-bridge#usage for more information. Stdout: ]]> Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: 3600 Stdout: Stdout: Stdout: Stdout: 3600 Stdout: Stdout: Stdout: 60 Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: system Stdout: query_log
Stdout: Stdout: toYYYYMM(event_date) Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: 7500 Stdout: Stdout: 1048576 Stdout: Stdout: 8192 Stdout: Stdout: 524288 Stdout: Stdout: false Stdout: Stdout: Stdout: Stdout:
Stdout: Stdout: Stdout: Stdout: system Stdout: trace_log
Stdout: Stdout: toYYYYMM(event_date) Stdout: 7500 Stdout: 1048576 Stdout: 8192 Stdout: 524288 Stdout: Stdout: false Stdout: true Stdout:
Stdout: Stdout: Stdout: Stdout: system Stdout: query_thread_log
Stdout: toYYYYMM(event_date) Stdout: 7500 Stdout: 1048576 Stdout: 8192 Stdout: 524288 Stdout: false Stdout:
Stdout: Stdout: Stdout: Stdout: system Stdout: query_views_log
Stdout: toYYYYMM(event_date) Stdout: 7500 Stdout:
Stdout: Stdout: Stdout: Stdout: system Stdout: part_log
Stdout: toYYYYMM(event_date) Stdout: 7500 Stdout: 1048576 Stdout: 8192 Stdout: 524288 Stdout: false Stdout:
Stdout: Stdout: Stdout: Stdout: system Stdout: text_log
Stdout: 7500 Stdout: 1048576 Stdout: 8192 Stdout: 524288 Stdout: false Stdout: trace Stdout:
Stdout: Stdout: Stdout: Stdout: system Stdout: metric_log
Stdout: 7500 Stdout: 1048576 Stdout: 8192 Stdout: 524288 Stdout: 1000 Stdout: false Stdout:
Stdout: Stdout: Stdout: Stdout: system Stdout: latency_log
Stdout: 7500 Stdout: 1048576 Stdout: 8192 Stdout: 524288 Stdout: 1000 Stdout: false Stdout:
Stdout: Stdout: Stdout: Stdout: system Stdout: error_log
Stdout: 7500 Stdout: 1048576 Stdout: 8192 run container_id:roottestattachtablenormalizer-gw6-node-1 detach:False nothrow:False cmd: ['bash', '-c', "sed -i 's/substring/substr/g' /var/lib/clickhouse/metadata/default/file.sql"] Stdout: 524288 Command:[docker exec -u root roottestattachtablenormalizer-gw6-node-1 bash -c sed -i 's/substring/substr/g' /var/lib/clickhouse/metadata/default/file.sql] Stdout: 1000 Stdout: false Stdout:
Stdout: Stdout: Stdout: Stdout: system Stdout: query_metric_log
Stdout: 7500 Stdout: 1048576 Stdout: 8192 Stdout: 524288 Stdout: 1000 Stdout: false Stdout:
Stdout: Stdout: Stdout: Stdout: system Stdout: asynchronous_metric_log
Stdout: 7000 Stdout: 1048576 Stdout: 8192 Stdout: 524288 Stdout: false Stdout:
Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: engine MergeTree Stdout: partition by toYYYYMM(finish_date) Stdout: order by (finish_date, finish_time_us, trace_id) Stdout: Stdout: system Stdout: opentelemetry_span_log
Stdout: 7500 Stdout: 1048576 Stdout: 8192 Stdout: 524288 Stdout: false Stdout:
Stdout: Stdout: Stdout: Stdout: Stdout: system Stdout: crash_log
Stdout: Stdout: Stdout: 1000 Stdout: 1024 Stdout: 1024 Stdout: 512 Stdout: true Stdout:
Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: system Stdout: processors_profile_log
Stdout: Stdout: toYYYYMM(event_date) Stdout: 7500 Stdout: 1048576 Stdout: 8192 Stdout: 524288 Stdout: false Stdout: event_date + INTERVAL 30 DAY DELETE Stdout:
Stdout: Stdout: Stdout: Stdout: system Stdout: asynchronous_insert_log
Stdout: Stdout: 7500 Stdout: 1048576 Stdout: 8192 Stdout: 524288 Stdout: false Stdout: event_date Stdout: event_date + INTERVAL 3 DAY Stdout:
Stdout: Stdout: Stdout: Stdout: system Stdout: backup_log
Stdout: toYYYYMM(event_date) Stdout: 7500 Stdout:
Stdout: Stdout: Stdout: Stdout: system Stdout: s3queue_log
Stdout: toYYYYMM(event_date) Stdout: 7500 Stdout:
Stdout: Stdout: Stdout: Stdout: system Stdout: blob_storage_log
Stdout: toYYYYMM(event_date) Stdout: 7500 Stdout: event_date + INTERVAL 30 DAY Stdout:
Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: *_dictionary.*ml Stdout: Stdout: Stdout: true Stdout: Stdout: Stdout: true Stdout: Stdout: Stdout: *_function.*ml Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: /clickhouse/task_queue/ddl Stdout: Stdout: /clickhouse/task_queue/replicas Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: click_cost Stdout: any Stdout: Stdout: 0 Stdout: 3600 Stdout: Stdout: Stdout: 86400 Stdout: 60 Stdout: Stdout: Stdout: Stdout: max Stdout: Stdout: 0 Stdout: 60 Stdout: Stdout: Stdout: 3600 Stdout: 300 Stdout: Stdout: Stdout: 86400 Stdout: 3600 Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: /var/lib/clickhouse/format_schemas/ Stdout: Stdout: Stdout: /usr/share/clickhouse/protos/ Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: false Stdout: Stdout: false Stdout: Stdout: Stdout: https://6f33034cfe684dd7a3ab9875e57b1c8d@o388870.ingest.sentry.io/5226277 Stdout: Stdout: false Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: backups Stdout: Stdout: Stdout: true Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: node Stdout: Stdout: Stdout: Stdout: Stdout: 1 Stdout: Stdout: Stdout: Stdout: Etc/UTC Stdout: 0.0.0.0 Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: users.xml Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: 1 Stdout: 1 Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: Stdout: 40000000000 Stdout: Stdout:
[gw9] PASSED test_config_hide_in_preprocessed/test.py::test_hide_in_preprocessed Command:[docker compose --env-file /ClickHouse/tests/integration/test_config_hide_in_preprocessed/_instances-0-gw9/.env --project-name roottestconfighideinpreprocessed-gw9 --file /ClickHouse/tests/integration/test_config_hide_in_preprocessed/_instances-0-gw9/node/docker-compose.yml stop --timeout 20] Executing query select dictGetFloat64('MySQL_complex_key_cache_', 'Float64_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictHas('RedisSimple_flat_Float32_', toUInt64(2)) on node run container_id:roottestattachtablenormalizer-gw6-node-1 detach:False nothrow:True cmd: ['bash', '-c', 'ps -C clickhouse'] Command:[docker exec -u root roottestattachtablenormalizer-gw6-node-1 bash -c ps -C clickhouse] Stdout: PID TTY TIME CMD Stdout: 8 ? 00:00:02 clickhouse run container_id:roottestattachtablenormalizer-gw6-node-1 detach:False nothrow:False cmd: ['bash', '-c', 'pkill -9 clickhouse'] Command:[docker exec -u root roottestattachtablenormalizer-gw6-node-1 bash -c pkill -9 clickhouse] run container_id:roottestattachtablenormalizer-gw6-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestattachtablenormalizer-gw6-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select dictHas('LocalClickHouse_complex_key_direct_', tuple(toUInt64(1),toString('world'))) on local_node Stdout:8 Executing query SELECT * FROM tbl ORDER BY x on node1 Executing query select count() from system.text_log_7_test on node2 Executing query select dictHas('MySQL_complex_key_cache_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGetOrDefault('RedisSimple_flat_Float32_', 'Float32_', toUInt64(9999999), toFloat32(555.11)) on node Executing query select dictGetOrDefault('LocalClickHouse_complex_key_direct_', 'Int64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt64(-88)) on local_node run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query SELECT * FROM tbl ORDER BY x on node2 Stdout:4111 Executing query select count() from system.query_log_13_test on node2 Executing query select dictGetFloat32OrDefault('RedisSimple_flat_Float32_', 'Float32_', toUInt64(9999999), toFloat32(555.11)) on node Executing query select dictGetOrDefault('MySQL_complex_key_cache_', 'Float64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat64(777.11)) on node Executing query select dictGetInt64OrDefault('LocalClickHouse_complex_key_direct_', 'Int64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt64(-88)) on local_node Executing query system reload dictionary RedisSimple_hashed_Float32_ on node Executing query DROP TABLE IF EXISTS tbl ON CLUSTER 'cluster3' SYNC on node1 [gw2] PASSED test_backup_restore_on_cluster/test.py::test_keeper_value_max_size Executing query select count() from system.asynchronous_loader where job ilike '%_log_%_test' and execution_pool = 'BackgroundLoad' on node2 Executing query select dictGetFloat64OrDefault('MySQL_complex_key_cache_', 'Float64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat64(777.11)) on node Executing query select dictGet('LocalClickHouse_complex_key_direct_', 'UUID_', tuple(toUInt64(1),toString('world'))) on local_node Executing query select dictGet('RedisSimple_hashed_Float32_', 'Float32_', toUInt64(1)) on node run container_id:roottestattachtablenormalizer-gw6-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestattachtablenormalizer-gw6-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query system flush logs on node2 run container_id:roottestattachtablenormalizer-gw6-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestattachtablenormalizer-gw6-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] [gw3] PASSED test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_complex[complex_key_cache] Executing query system reload dictionaries on node test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_complex[complex_key_direct] No clickhouse process running. Start new one. http://localhost:None "POST /v1.46/containers/roottestattachtablenormalizer-gw6-node-1/exec HTTP/1.1" 201 74 http://localhost:None "POST /v1.46/exec/bf7be1280ccf65b1954d23ab7b76c2112316d93f6f2a4faabcdebb30d9fc547a/start HTTP/1.1" 200 0 http://localhost:None "GET /v1.46/exec/bf7be1280ccf65b1954d23ab7b76c2112316d93f6f2a4faabcdebb30d9fc547a/json HTTP/1.1" 200 586 Executing query DROP TABLE IF EXISTS tbl2 ON CLUSTER 'cluster3' SYNC on node1 Executing query select dictGetUUID('LocalClickHouse_complex_key_direct_', 'UUID_', tuple(toUInt64(1),toString('world'))) on local_node Executing query select dictGetFloat32('RedisSimple_hashed_Float32_', 'Float32_', toUInt64(1)) on node Executing query select dictGet('MySQL_complex_key_direct_', 'UInt8_', tuple(toUInt64(1),toString('world'))) on node run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query DROP DATABASE IF EXISTS mydb ON CLUSTER 'cluster3' SYNC on node1 run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] No clickhouse process running. Start new one. http://localhost:None "POST /v1.46/containers/roottestdatabasebackup-gw1-instance-1/exec HTTP/1.1" 201 74 http://localhost:None "POST /v1.46/exec/bda65f6dd5cd4a3332738a69786fb65b5d8323c594ac1b28820f3450ad83bea3/start HTTP/1.1" 200 0 http://localhost:None "GET /v1.46/exec/bda65f6dd5cd4a3332738a69786fb65b5d8323c594ac1b28820f3450ad83bea3/json HTTP/1.1" 200 586 Executing query select dictHas('LocalClickHouse_complex_key_direct_', tuple(toUInt64(1),toString('world'))) on local_node Executing query select dictHas('RedisSimple_hashed_Float32_', toUInt64(1)) on node Executing query select dictGetUInt8('MySQL_complex_key_direct_', 'UInt8_', tuple(toUInt64(1),toString('world'))) on node Stderr: Container roottestconfighideinpreprocessed-gw9-node-1 Stopping Stderr: Container roottestconfighideinpreprocessed-gw9-node-1 Stopped Command:[bash -c [ -f /ClickHouse/tests/integration/test_config_hide_in_preprocessed/_instances-0-gw9/node/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_config_hide_in_preprocessed/_instances-0-gw9/node/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[docker compose --env-file /ClickHouse/tests/integration/test_config_hide_in_preprocessed/_instances-0-gw9/.env --project-name roottestconfighideinpreprocessed-gw9 --file /ClickHouse/tests/integration/test_config_hide_in_preprocessed/_instances-0-gw9/node/docker-compose.yml down --volumes] Executing query select dictHas('MySQL_complex_key_direct_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGetOrDefault('RedisSimple_hashed_Float32_', 'Float32_', toUInt64(9999999), toFloat32(555.11)) on node Executing query select dictGetOrDefault('LocalClickHouse_complex_key_direct_', 'UUID_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUUID('550e8400-0000-0000-0000-000000000000')) on local_node Executing query DROP DATABASE IF EXISTS mydb2 ON CLUSTER 'cluster3' SYNC on node1 Executing query select count() from system.tables where database = 'system' and name in ['query_log', 'text_log'] on node2 run container_id:roottestattachtablenormalizer-gw6-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestattachtablenormalizer-gw6-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:739 Clickhouse process running. run container_id:roottestattachtablenormalizer-gw6-node-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestattachtablenormalizer-gw6-node-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:739 Executing query select 20 on node Executing query select dictGetUUIDOrDefault('LocalClickHouse_complex_key_direct_', 'UUID_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUUID('550e8400-0000-0000-0000-000000000000')) on local_node Executing query rename table system.text_log to system.text_log_21_test on node2 Executing query select dictGetOrDefault('MySQL_complex_key_direct_', 'UInt8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt8(55)) on node Executing query select dictGetFloat32OrDefault('RedisSimple_hashed_Float32_', 'Float32_', toUInt64(9999999), toFloat32(555.11)) on node Executing query DROP USER IF EXISTS u1, u2 ON CLUSTER 'cluster3' on node1 run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query rename table system.query_log to system.query_log_21_test on node2 Stdout:4938 Clickhouse process running. run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select dictGetUInt8OrDefault('MySQL_complex_key_direct_', 'UInt8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt8(55)) on node Stdout:4938 Executing query select dictGet('LocalClickHouse_complex_key_direct_', 'Date_', tuple(toUInt64(1),toString('world'))) on local_node Executing query select 20 on instance Executing query select dictGet('RedisSimple_hashed_Float32_', 'Float32_', toUInt64(2)) on node Stderr: Container roottestconfighideinpreprocessed-gw9-node-1 Stopping Stderr: Container roottestconfighideinpreprocessed-gw9-node-1 Stopped Stderr: Container roottestconfighideinpreprocessed-gw9-node-1 Removing Stderr: Container roottestconfighideinpreprocessed-gw9-node-1 Removed Stderr: Network roottestconfighideinpreprocessed-gw9_default Removing Stderr: Network roottestconfighideinpreprocessed-gw9_default Removed Cleanup called Docker networks for project roottestconfighideinpreprocessed-gw9 are NETWORK ID NAME DRIVER SCOPE test_backup_restore_on_cluster/test.py::test_mutation Executing query CREATE TABLE tbl ON CLUSTER 'cluster' (x UInt8, y String) ENGINE=ReplicatedMergeTree('/clickhouse/tables/tbl/', '{replica}')ORDER BY tuple() on node1 Docker containers for project roottestconfighideinpreprocessed-gw9 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Executing query system flush logs on node2 Docker volumes for project roottestconfighideinpreprocessed-gw9 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestconfighideinpreprocessed-gw9-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} No running containers for project: roottestconfighideinpreprocessed-gw9 Trying to prune unused networks... Trying to prune unused images... Command:[docker image prune -f] Executing query select 20 on node Executing query select dictGet('MySQL_complex_key_direct_', 'UInt16_', tuple(toUInt64(1),toString('world'))) on node Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Stdout:5 Command:[docker volume prune -f] Stdout:Total reclaimed space: 0B Volumes pruned: 5 Executing query select dictGetFloat32('RedisSimple_hashed_Float32_', 'Float32_', toUInt64(2)) on node Executing query select dictGetDate('LocalClickHouse_complex_key_direct_', 'Date_', tuple(toUInt64(1),toString('world'))) on local_node Executing query select dictGetUInt16('MySQL_complex_key_direct_', 'UInt16_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictHas('RedisSimple_hashed_Float32_', toUInt64(2)) on node Executing query select dictHas('LocalClickHouse_complex_key_direct_', tuple(toUInt64(1),toString('world'))) on local_node Executing query select 20 on instance Executing query INSERT INTO tbl SELECT number, toString(number) FROM numbers(5) on node1 Executing query select dictHas('MySQL_complex_key_direct_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGetOrDefault('RedisSimple_hashed_Float32_', 'Float32_', toUInt64(9999999), toFloat32(555.11)) on node Executing query select 20 on node Executing query select dictGetOrDefault('LocalClickHouse_complex_key_direct_', 'Date_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDate('2018-12-30')) on local_node Executing query INSERT INTO tbl SELECT number, toString(number) FROM numbers(5, 5) on node2 Executing query select dictGetFloat32OrDefault('RedisSimple_hashed_Float32_', 'Float32_', toUInt64(9999999), toFloat32(555.11)) on node Executing query select dictGetOrDefault('MySQL_complex_key_direct_', 'UInt16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt16(66)) on node Executing query select count() from system.tables where database = 'system' and name in ['query_log', 'text_log'] on node2 Command:[docker compose --env-file /ClickHouse/tests/integration/test_attach_table_normalizer/_instances-0-gw6/.env --project-name roottestattachtablenormalizer-gw6 --file /ClickHouse/tests/integration/test_attach_table_normalizer/_instances-0-gw6/node/docker-compose.yml stop --timeout 20] [gw6] PASSED test_attach_table_normalizer/test.py::test_attach_substr_restart Executing query select dictGetDateOrDefault('LocalClickHouse_complex_key_direct_', 'Date_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDate('2018-12-30')) on local_node Executing query INSERT INTO tbl SELECT number, toString(number) FROM numbers(10, 5) on node1 Executing query select 20 on instance Executing query system reload dictionary RedisSimple_cache_Float32_ on node Executing query select dictGetUInt16OrDefault('MySQL_complex_key_direct_', 'UInt16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt16(66)) on node Executing query rename table system.text_log to system.text_log_22_test on node2 Executing query select dictGet('LocalClickHouse_complex_key_direct_', 'DateTime_', tuple(toUInt64(1),toString('world'))) on local_node Stderr: Container roottestattachtablenormalizer-gw6-node-1 Stopping Stderr: Container roottestattachtablenormalizer-gw6-node-1 Stopped Command:[bash -c [ -f /ClickHouse/tests/integration/test_attach_table_normalizer/_instances-0-gw6/node/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_attach_table_normalizer/_instances-0-gw6/node/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[docker compose --env-file /ClickHouse/tests/integration/test_attach_table_normalizer/_instances-0-gw6/.env --project-name roottestattachtablenormalizer-gw6 --file /ClickHouse/tests/integration/test_attach_table_normalizer/_instances-0-gw6/node/docker-compose.yml down --volumes] Executing query select dictGet('RedisSimple_cache_Float32_', 'Float32_', toUInt64(1)) on node Executing query rename table system.query_log to system.query_log_22_test on node2 Executing query ALTER TABLE tbl UPDATE x=x+1 WHERE 1 on node1 Executing query select dictGet('MySQL_complex_key_direct_', 'UInt32_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGetDateTime('LocalClickHouse_complex_key_direct_', 'DateTime_', tuple(toUInt64(1),toString('world'))) on local_node Executing query system flush logs on node2 Executing query ALTER TABLE tbl UPDATE x=x+1+sleep(3) WHERE 1 on node1 Executing query select dictGetUInt32('MySQL_complex_key_direct_', 'UInt32_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGetFloat32('RedisSimple_cache_Float32_', 'Float32_', toUInt64(1)) on node Executing query select 20 on instance Executing query ALTER TABLE tbl UPDATE x=x+1+sleep(3) WHERE 1 on node1 Executing query select dictHas('LocalClickHouse_complex_key_direct_', tuple(toUInt64(1),toString('world'))) on local_node Executing query select dictHas('RedisSimple_cache_Float32_', toUInt64(1)) on node Stderr: Container roottestattachtablenormalizer-gw6-node-1 Stopping Stderr: Container roottestattachtablenormalizer-gw6-node-1 Stopped Stderr: Container roottestattachtablenormalizer-gw6-node-1 Removing Stderr: Container roottestattachtablenormalizer-gw6-node-1 Removed Stderr: Network roottestattachtablenormalizer-gw6_default Removing Stderr: Network roottestattachtablenormalizer-gw6_default Removed Cleanup called Docker networks for project roottestattachtablenormalizer-gw6 are NETWORK ID NAME DRIVER SCOPE Executing query select dictHas('MySQL_complex_key_direct_', tuple(toUInt64(1),toString('world'))) on node Docker containers for project roottestattachtablenormalizer-gw6 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestattachtablenormalizer-gw6 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestattachtablenormalizer-gw6-.*-1$' --format '{{.ID}}:{{.Names}}'] Executing query SELECT id, value FROM test_table_backup.test_table on instance Unstopped containers: {} No running containers for project: roottestattachtablenormalizer-gw6 Trying to prune unused networks... Trying to prune unused images... Command:[docker image prune -f] Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Stdout:5 Command:[docker volume prune -f] Executing query BACKUP TABLE tbl ON CLUSTER 'cluster' TO Disk('backups', '10') on node1 Stdout:Total reclaimed space: 0B Volumes pruned: 5 Executing query select dictGetOrDefault('LocalClickHouse_complex_key_direct_', 'DateTime_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDateTime('2018-12-30 00:00:00')) on local_node Executing query select dictGetOrDefault('RedisSimple_cache_Float32_', 'Float32_', toUInt64(9999999), toFloat32(555.11)) on node Executing query DROP DATABASE test_table_backup on instance Executing query select dictGetOrDefault('MySQL_complex_key_direct_', 'UInt32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt32(77)) on node Executing query select dictGetFloat32OrDefault('RedisSimple_cache_Float32_', 'Float32_', toUInt64(9999999), toFloat32(555.11)) on node Executing query select dictGetDateTimeOrDefault('LocalClickHouse_complex_key_direct_', 'DateTime_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDateTime('2018-12-30 00:00:00')) on local_node Executing query DROP DATABASE test_database on instance Executing query select count() from system.tables where database = 'system' and name in ['query_log', 'text_log'] on node2 Executing query select dictGetUInt32OrDefault('MySQL_complex_key_direct_', 'UInt32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt32(77)) on node run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', 'rm -rf /backups/'] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c rm -rf /backups/] Executing query select dictGet('RedisSimple_cache_Float32_', 'Float32_', toUInt64(2)) on node Executing query select dictGet('LocalClickHouse_complex_key_direct_', 'String_', tuple(toUInt64(1),toString('world'))) on local_node run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', 'rm -rf /local_plain/'] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c rm -rf /local_plain/] Executing query DROP TABLE tbl ON CLUSTER 'cluster' SYNC on node1 http://172.16.8.4:9001 "GET /root?delimiter=&encoding-type=url&list-type=2&max-keys=1000&prefix= HTTP/1.1" 200 0 [gw1] PASSED test_database_backup/test.py::test_database_backup_table[Disk('backup_disk_object_storage_local_plain', 'test_table_backup')] test_database_backup/test.py::test_database_backup_table[Disk('backup_disk_s3_plain', 'test_table_backup')] run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', 'rm -rf /backups/'] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c rm -rf /backups/] Executing query select dictGet('MySQL_complex_key_direct_', 'UInt64_', tuple(toUInt64(1),toString('world'))) on node run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', 'rm -rf /local_plain/'] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c rm -rf /local_plain/] http://172.16.8.4:9001 "GET /root?delimiter=&encoding-type=url&list-type=2&max-keys=1000&prefix= HTTP/1.1" 200 0 Executing query DROP DATABASE IF EXISTS test_database; DROP DATABASE IF EXISTS test_table_backup; CREATE DATABASE test_database; CREATE TABLE test_database.test_table (id UInt64, value String) ENGINE=MergeTree ORDER BY id; INSERT INTO test_database.test_table VALUES (0, 'test_database.test_table'); BACKUP TABLE test_database.test_table TO Disk('backup_disk_s3_plain', 'test_table_backup'); CREATE DATABASE test_table_backup ENGINE = Backup('test_database', Disk('backup_disk_s3_plain', 'test_table_backup')); on instance Executing query rename table system.text_log to system.text_log_23_test on node2 Executing query select dictGetFloat32('RedisSimple_cache_Float32_', 'Float32_', toUInt64(2)) on node Executing query select dictGetString('LocalClickHouse_complex_key_direct_', 'String_', tuple(toUInt64(1),toString('world'))) on local_node Executing query select dictGetUInt64('MySQL_complex_key_direct_', 'UInt64_', tuple(toUInt64(1),toString('world'))) on node Executing query rename table system.query_log to system.query_log_23_test on node2 Executing query select dictHas('RedisSimple_cache_Float32_', toUInt64(2)) on node Executing query select dictHas('LocalClickHouse_complex_key_direct_', tuple(toUInt64(1),toString('world'))) on local_node Executing query system flush logs on node2 Executing query SELECT id, value FROM test_table_backup.test_table on instance Executing query select dictHas('MySQL_complex_key_direct_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGetOrDefault('RedisSimple_cache_Float32_', 'Float32_', toUInt64(9999999), toFloat32(555.11)) on node run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:True cmd: ['bash', '-c', 'ps -C clickhouse'] Command:[docker exec -u root roottestdatabasebackup-gw1-instance-1 bash -c ps -C clickhouse] Executing query select dictGetOrDefault('MySQL_complex_key_direct_', 'UInt64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt64(88)) on node Executing query select dictGetOrDefault('LocalClickHouse_complex_key_direct_', 'String_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toString('hi')) on local_node Stdout: PID TTY TIME CMD Stdout: 4938 ? 00:00:07 clickhouse run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', 'pkill clickhouse'] Command:[docker exec -u root roottestdatabasebackup-gw1-instance-1 bash -c pkill clickhouse] run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select dictGetFloat32OrDefault('RedisSimple_cache_Float32_', 'Float32_', toUInt64(9999999), toFloat32(555.11)) on node Stdout:4938 Executing query select dictGetUInt64OrDefault('MySQL_complex_key_direct_', 'UInt64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt64(88)) on node Executing query select dictGetStringOrDefault('LocalClickHouse_complex_key_direct_', 'String_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toString('hi')) on local_node Executing query system reload dictionary RedisSimple_direct_Float32_ on node Executing query select dictGet('MySQL_complex_key_direct_', 'Int8_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGet('RedisSimple_direct_Float32_', 'Float32_', toUInt64(1)) on node Executing query select dictGet('LocalClickHouse_complex_key_direct_', 'Float32_', tuple(toUInt64(1),toString('world'))) on local_node Executing query select count() from system.tables where database = 'system' and name in ['query_log', 'text_log'] on node2 Executing query select dictGetInt8('MySQL_complex_key_direct_', 'Int8_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGetFloat32('RedisSimple_direct_Float32_', 'Float32_', toUInt64(1)) on node Executing query rename table system.text_log to system.text_log_24_test on node2 Executing query select dictGetFloat32('LocalClickHouse_complex_key_direct_', 'Float32_', tuple(toUInt64(1),toString('world'))) on local_node run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:4938 Executing query select dictHas('MySQL_complex_key_direct_', tuple(toUInt64(1),toString('world'))) on node Executing query rename table system.query_log to system.query_log_24_test on node2 Executing query select dictHas('RedisSimple_direct_Float32_', toUInt64(1)) on node Executing query select dictHas('LocalClickHouse_complex_key_direct_', tuple(toUInt64(1),toString('world'))) on local_node Executing query select dictGetOrDefault('MySQL_complex_key_direct_', 'Int8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt8(-55)) on node Executing query system flush logs on node2 Executing query select dictGetOrDefault('RedisSimple_direct_Float32_', 'Float32_', toUInt64(9999999), toFloat32(555.11)) on node Executing query RESTORE TABLE tbl ON CLUSTER 'cluster' FROM Disk('backups', '10') on node1 Executing query select dictGetOrDefault('LocalClickHouse_complex_key_direct_', 'Float32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat32(555.11)) on local_node Executing query select dictGetInt8OrDefault('MySQL_complex_key_direct_', 'Int8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt8(-55)) on node Executing query select dictGetFloat32OrDefault('RedisSimple_direct_Float32_', 'Float32_', toUInt64(9999999), toFloat32(555.11)) on node Executing query select dictGetFloat32OrDefault('LocalClickHouse_complex_key_direct_', 'Float32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat32(555.11)) on local_node Executing query select dictGet('RedisSimple_direct_Float32_', 'Float32_', toUInt64(2)) on node Executing query select dictGet('MySQL_complex_key_direct_', 'Int16_', tuple(toUInt64(1),toString('world'))) on node run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query DROP TABLE IF EXISTS tbl ON CLUSTER 'cluster3' SYNC on node1 [gw2] PASSED test_backup_restore_on_cluster/test.py::test_mutation Stdout:4938 Executing query select dictGet('LocalClickHouse_complex_key_direct_', 'Float64_', tuple(toUInt64(1),toString('world'))) on local_node Executing query select dictGetFloat32('RedisSimple_direct_Float32_', 'Float32_', toUInt64(2)) on node Executing query select dictGetInt16('MySQL_complex_key_direct_', 'Int16_', tuple(toUInt64(1),toString('world'))) on node Executing query select count() from system.tables where database = 'system' and name in ['query_log', 'text_log'] on node2 Executing query select dictGetFloat64('LocalClickHouse_complex_key_direct_', 'Float64_', tuple(toUInt64(1),toString('world'))) on local_node Executing query select dictHas('RedisSimple_direct_Float32_', toUInt64(2)) on node Executing query DROP TABLE IF EXISTS tbl2 ON CLUSTER 'cluster3' SYNC on node1 Executing query select dictHas('MySQL_complex_key_direct_', tuple(toUInt64(1),toString('world'))) on node Executing query rename table system.text_log to system.text_log_25_test on node2 Executing query select dictGetOrDefault('RedisSimple_direct_Float32_', 'Float32_', toUInt64(9999999), toFloat32(555.11)) on node Executing query select dictHas('LocalClickHouse_complex_key_direct_', tuple(toUInt64(1),toString('world'))) on local_node Executing query select dictGetOrDefault('MySQL_complex_key_direct_', 'Int16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt16(-66)) on node Executing query rename table system.query_log to system.query_log_25_test on node2 Executing query DROP DATABASE IF EXISTS mydb ON CLUSTER 'cluster3' SYNC on node1 Executing query select dictGetFloat32OrDefault('RedisSimple_direct_Float32_', 'Float32_', toUInt64(9999999), toFloat32(555.11)) on node Executing query system flush logs on node2 Executing query select dictGetOrDefault('LocalClickHouse_complex_key_direct_', 'Float64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat64(777.11)) on local_node run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:4938 Executing query select dictGetInt16OrDefault('MySQL_complex_key_direct_', 'Int16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt16(-66)) on node Executing query DROP DATABASE IF EXISTS mydb2 ON CLUSTER 'cluster3' SYNC on node1 Executing query system reload dictionary RedisHash_complex_key_hashed_Float32_ on node Executing query select dictGetFloat64OrDefault('LocalClickHouse_complex_key_direct_', 'Float64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat64(777.11)) on local_node Executing query select dictGet('MySQL_complex_key_direct_', 'Int32_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGet('RedisHash_complex_key_hashed_Float32_', 'Float32_', tuple(toUInt64(1),toString('world'))) on node Executing query DROP USER IF EXISTS u1, u2 ON CLUSTER 'cluster3' on node1 Stderr: Container roottestclusterallreplicas-gw4-node1-1 Stopping Stderr: Container roottestclusterallreplicas-gw4-node2-1 Stopping Stderr: Container roottestclusterallreplicas-gw4-node2-1 Stopped Stderr: Container roottestclusterallreplicas-gw4-node1-1 Stopped Command:[bash -c [ -f /ClickHouse/tests/integration/test_cluster_all_replicas/_instances-0-gw4/node1/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_cluster_all_replicas/_instances-0-gw4/node1/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[bash -c [ -f /ClickHouse/tests/integration/test_cluster_all_replicas/_instances-0-gw4/node2/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_cluster_all_replicas/_instances-0-gw4/node2/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Executing query select dictGet('LocalClickHouse_complex_key_direct_', 'UInt8_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Command:[docker compose --env-file /ClickHouse/tests/integration/test_cluster_all_replicas/_instances-0-gw4/.env --project-name roottestclusterallreplicas-gw4 --file /ClickHouse/tests/integration/test_cluster_all_replicas/_instances-0-gw4/node1/docker-compose.yml --file /ClickHouse/tests/integration/test_cluster_all_replicas/_instances-0-gw4/node2/docker-compose.yml down --volumes] Executing query select dictGetInt32('MySQL_complex_key_direct_', 'Int32_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGetFloat32('RedisHash_complex_key_hashed_Float32_', 'Float32_', tuple(toUInt64(1),toString('world'))) on node test_backup_restore_on_cluster/test.py::test_projection Executing query CREATE TABLE tbl ON CLUSTER 'cluster' (x UInt32, y String) ENGINE=ReplicatedMergeTree('/clickhouse/tables/tbl/', '{replica}') ORDER BY y PARTITION BY x%10 on node1 Executing query select count() from system.tables where database = 'system' and name in ['query_log', 'text_log'] on node2 Executing query select dictHas('RedisHash_complex_key_hashed_Float32_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictHas('MySQL_complex_key_direct_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGetUInt8('LocalClickHouse_complex_key_direct_', 'UInt8_', tuple(toUInt64(2),toString('qwerty2'))) on local_node run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:4938 Executing query rename table system.text_log to system.text_log_26_test on node2 Stderr: Container roottestclusterallreplicas-gw4-node1-1 Stopping Stderr: Container roottestclusterallreplicas-gw4-node2-1 Stopping Stderr: Container roottestclusterallreplicas-gw4-node1-1 Stopped Stderr: Container roottestclusterallreplicas-gw4-node1-1 Removing Stderr: Container roottestclusterallreplicas-gw4-node2-1 Stopped Stderr: Container roottestclusterallreplicas-gw4-node2-1 Removing Stderr: Container roottestclusterallreplicas-gw4-node1-1 Removed Stderr: Container roottestclusterallreplicas-gw4-node2-1 Removed Stderr: Network roottestclusterallreplicas-gw4_default Removing Stderr: Network roottestclusterallreplicas-gw4_default Removed Cleanup called Executing query select dictGetOrDefault('MySQL_complex_key_direct_', 'Int32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt32(-77)) on node Docker networks for project roottestclusterallreplicas-gw4 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestclusterallreplicas-gw4 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Executing query select dictGetOrDefault('RedisHash_complex_key_hashed_Float32_', 'Float32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat32(555.11)) on node Docker volumes for project roottestclusterallreplicas-gw4 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestclusterallreplicas-gw4-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} No running containers for project: roottestclusterallreplicas-gw4 Trying to prune unused networks... Trying to prune unused images... Command:[docker image prune -f] Executing query INSERT INTO tbl SELECT number, toString(number) FROM numbers(3) on node1 Executing query select dictHas('LocalClickHouse_complex_key_direct_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Stdout:5 Command:[docker volume prune -f] Stdout:Total reclaimed space: 0B Volumes pruned: 5 test_aggregation_memory_efficient/test.py::test_remote Running tests in /ClickHouse/tests/integration/test_aggregation_memory_efficient/test.py Cluster start called. is_up=False Docker networks for project roottestaggregationmemoryefficient-gw4 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestaggregationmemoryefficient-gw4 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Executing query rename table system.query_log to system.query_log_26_test on node2 Docker volumes for project roottestaggregationmemoryefficient-gw4 are DRIVER VOLUME NAME Cleanup called Docker networks for project roottestaggregationmemoryefficient-gw4 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestaggregationmemoryefficient-gw4 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestaggregationmemoryefficient-gw4 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestaggregationmemoryefficient-gw4-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} No running containers for project: roottestaggregationmemoryefficient-gw4 Trying to prune unused networks... Executing query select dictGetInt32OrDefault('MySQL_complex_key_direct_', 'Int32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt32(-77)) on node Trying to prune unused images... Command:[docker image prune -f] Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Stdout:5 Command:[docker volume prune -f] Stdout:Total reclaimed space: 0B Volumes pruned: 5 Setup directory for instance: node1 Create directory for configuration generated in this helper Create directory for common tests configuration Copy common configuration from helpers Generate and write macros file Copy custom test config files [] to /ClickHouse/tests/integration/test_aggregation_memory_efficient/_instances-0-gw4/node1/configs/config.d Setup database dir /ClickHouse/tests/integration/test_aggregation_memory_efficient/_instances-0-gw4/node1/database Setup logs dir /ClickHouse/tests/integration/test_aggregation_memory_efficient/_instances-0-gw4/node1/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Setup directory for instance: node2 Create directory for configuration generated in this helper Create directory for common tests configuration Copy common configuration from helpers Generate and write macros file Copy custom test config files [] to /ClickHouse/tests/integration/test_aggregation_memory_efficient/_instances-0-gw4/node2/configs/config.d Setup database dir /ClickHouse/tests/integration/test_aggregation_memory_efficient/_instances-0-gw4/node2/database Setup logs dir /ClickHouse/tests/integration/test_aggregation_memory_efficient/_instances-0-gw4/node2/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw'} stored in /ClickHouse/tests/integration/test_aggregation_memory_efficient/_instances-0-gw4/.env Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found http://localhost:None "GET /version HTTP/1.1" 200 826 Command:[docker compose --env-file /ClickHouse/tests/integration/test_aggregation_memory_efficient/_instances-0-gw4/.env --project-name roottestaggregationmemoryefficient-gw4 --file /ClickHouse/tests/integration/test_aggregation_memory_efficient/_instances-0-gw4/node1/docker-compose.yml --file /ClickHouse/tests/integration/test_aggregation_memory_efficient/_instances-0-gw4/node2/docker-compose.yml pull] Executing query select dictGetFloat32OrDefault('RedisHash_complex_key_hashed_Float32_', 'Float32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat32(555.11)) on node Executing query system flush logs on node2 Executing query select dictGetOrDefault('LocalClickHouse_complex_key_direct_', 'UInt8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt8(55)) on local_node Executing query ALTER TABLE tbl ADD PROJECTION prjmax (SELECT MAX(x)) on node1 Executing query select dictGet('MySQL_complex_key_direct_', 'Int64_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGet('RedisHash_complex_key_hashed_Float32_', 'Float32_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGetUInt8OrDefault('LocalClickHouse_complex_key_direct_', 'UInt8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt8(55)) on local_node Executing query INSERT INTO tbl VALUES (100, 'a'), (101, 'b') on node1 Executing query select dictGetInt64('MySQL_complex_key_direct_', 'Int64_', tuple(toUInt64(1),toString('world'))) on node run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select dictGetFloat32('RedisHash_complex_key_hashed_Float32_', 'Float32_', tuple(toUInt64(2),toString('qwerty2'))) on node No clickhouse process running. Start new one. http://localhost:None "POST /v1.46/containers/roottestdatabasebackup-gw1-instance-1/exec HTTP/1.1" 201 74 http://localhost:None "POST /v1.46/exec/2c4286ea565161639d3fa0f200716a6863cdd99e888d50af3b5fe19c1658118d/start HTTP/1.1" 200 0 http://localhost:None "GET /v1.46/exec/2c4286ea565161639d3fa0f200716a6863cdd99e888d50af3b5fe19c1658118d/json HTTP/1.1" 200 586 Executing query SELECT count() FROM system.projection_parts WHERE database='default' AND table='tbl' AND name='prjmax' on node1 Executing query select dictGet('LocalClickHouse_complex_key_direct_', 'UInt16_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select dictHas('MySQL_complex_key_direct_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictHas('RedisHash_complex_key_hashed_Float32_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGetUInt16('LocalClickHouse_complex_key_direct_', 'UInt16_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query BACKUP TABLE tbl ON CLUSTER 'cluster' TO Disk('backups', '11') on node1 Executing query select count() from system.tables where database = 'system' and name in ['query_log', 'text_log'] on node2 Executing query select dictGetOrDefault('MySQL_complex_key_direct_', 'Int64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt64(-88)) on node Executing query select dictGetOrDefault('RedisHash_complex_key_hashed_Float32_', 'Float32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat32(555.11)) on node Connection dropped: socket connection error: None Executing query select dictHas('LocalClickHouse_complex_key_direct_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query rename table system.text_log to system.text_log_27_test on node2 Executing query select dictGetInt64OrDefault('MySQL_complex_key_direct_', 'Int64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt64(-88)) on node Executing query select dictGetFloat32OrDefault('RedisHash_complex_key_hashed_Float32_', 'Float32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat32(555.11)) on node Executing query rename table system.query_log to system.query_log_27_test on node2 Executing query select dictGetOrDefault('LocalClickHouse_complex_key_direct_', 'UInt16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt16(66)) on local_node Executing query select dictGet('MySQL_complex_key_direct_', 'UUID_', tuple(toUInt64(1),toString('world'))) on node run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query system flush logs on node2 Stdout:5784 Clickhouse process running. run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query system reload dictionary RedisHash_complex_key_cache_Float32_ on node Executing query DROP TABLE tbl ON CLUSTER 'cluster' SYNC on node1 Stdout:5784 Executing query select 20 on instance Executing query select dictGetUInt16OrDefault('LocalClickHouse_complex_key_direct_', 'UInt16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt16(66)) on local_node Executing query select dictGetUUID('MySQL_complex_key_direct_', 'UUID_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGet('RedisHash_complex_key_cache_Float32_', 'Float32_', tuple(toUInt64(1),toString('world'))) on node Executing query SELECT count() FROM system.projection_parts WHERE database='default' AND table='tbl' AND name='prjmax' on node1 Executing query select dictHas('MySQL_complex_key_direct_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGet('LocalClickHouse_complex_key_direct_', 'UInt32_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select dictGetFloat32('RedisHash_complex_key_cache_Float32_', 'Float32_', tuple(toUInt64(1),toString('world'))) on node Executing query select 20 on instance Executing query RESTORE TABLE tbl FROM Disk('backups', '11') on node1 Executing query select dictGetOrDefault('MySQL_complex_key_direct_', 'UUID_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUUID('550e8400-0000-0000-0000-000000000000')) on node Executing query select dictGetUInt32('LocalClickHouse_complex_key_direct_', 'UInt32_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select dictHas('RedisHash_complex_key_cache_Float32_', tuple(toUInt64(1),toString('world'))) on node Executing query select count() from system.tables where database = 'system' and name in ['query_log', 'text_log'] on node2 Executing query select dictGetUUIDOrDefault('MySQL_complex_key_direct_', 'UUID_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUUID('550e8400-0000-0000-0000-000000000000')) on node Executing query select dictGetOrDefault('RedisHash_complex_key_cache_Float32_', 'Float32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat32(555.11)) on node Executing query select dictHas('LocalClickHouse_complex_key_direct_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query SELECT * FROM tbl ORDER BY x on node1 Executing query rename table system.text_log to system.text_log_28_test on node2 Executing query select dictGet('MySQL_complex_key_direct_', 'Date_', tuple(toUInt64(1),toString('world'))) on node Executing query select 20 on instance Executing query select dictGetFloat32OrDefault('RedisHash_complex_key_cache_Float32_', 'Float32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat32(555.11)) on node Executing query select dictGetOrDefault('LocalClickHouse_complex_key_direct_', 'UInt32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt32(77)) on local_node Executing query rename table system.query_log to system.query_log_28_test on node2 Executing query SELECT count() FROM system.projection_parts WHERE database='default' AND table='tbl' AND name='prjmax' on node1 Executing query select dictGetDate('MySQL_complex_key_direct_', 'Date_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGet('RedisHash_complex_key_cache_Float32_', 'Float32_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query system flush logs on node2 Connection dropped: socket connection error: None Executing query select dictGetUInt32OrDefault('LocalClickHouse_complex_key_direct_', 'UInt32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt32(77)) on local_node Executing query DROP TABLE IF EXISTS tbl ON CLUSTER 'cluster3' SYNC on node1 [gw2] PASSED test_backup_restore_on_cluster/test.py::test_projection Executing query select dictHas('MySQL_complex_key_direct_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGetFloat32('RedisHash_complex_key_cache_Float32_', 'Float32_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select 20 on instance Executing query select dictGet('LocalClickHouse_complex_key_direct_', 'UInt64_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select dictGetOrDefault('MySQL_complex_key_direct_', 'Date_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDate('2018-12-30')) on node Executing query DROP TABLE IF EXISTS tbl2 ON CLUSTER 'cluster3' SYNC on node1 Executing query select dictHas('RedisHash_complex_key_cache_Float32_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query SELECT id, value FROM test_table_backup.test_table on instance Executing query select dictGetUInt64('LocalClickHouse_complex_key_direct_', 'UInt64_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select dictGetDateOrDefault('MySQL_complex_key_direct_', 'Date_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDate('2018-12-30')) on node Executing query select dictGetOrDefault('RedisHash_complex_key_cache_Float32_', 'Float32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat32(555.11)) on node Executing query DROP DATABASE IF EXISTS mydb ON CLUSTER 'cluster3' SYNC on node1 Executing query DROP DATABASE test_table_backup on instance Executing query select dictGet('MySQL_complex_key_direct_', 'DateTime_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictHas('LocalClickHouse_complex_key_direct_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select count() from system.tables where database = 'system' and name in ['query_log', 'text_log'] on node2 Executing query select dictGetFloat32OrDefault('RedisHash_complex_key_cache_Float32_', 'Float32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat32(555.11)) on node Executing query DROP DATABASE test_database on instance Executing query DROP DATABASE IF EXISTS mydb2 ON CLUSTER 'cluster3' SYNC on node1 Executing query select dictGetDateTime('MySQL_complex_key_direct_', 'DateTime_', tuple(toUInt64(1),toString('world'))) on node Executing query rename table system.text_log to system.text_log_29_test on node2 Executing query select dictGetOrDefault('LocalClickHouse_complex_key_direct_', 'UInt64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt64(88)) on local_node run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', 'rm -rf /backups/'] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c rm -rf /backups/] Executing query system reload dictionary RedisHash_complex_key_direct_Float32_ on node run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', 'rm -rf /local_plain/'] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c rm -rf /local_plain/] http://172.16.8.4:9001 "GET /root?delimiter=&encoding-type=url&list-type=2&max-keys=1000&prefix= HTTP/1.1" 200 0 http://172.16.8.4:9001 "DELETE /root/data/disks/disk_s3_plain/test_table_backup/.backup HTTP/1.1" 204 0 http://172.16.8.4:9001 "DELETE /root/data/disks/disk_s3_plain/test_table_backup/data/test_database/test_table/all_1_1_0/checksums.txt HTTP/1.1" 204 0 http://172.16.8.4:9001 "DELETE /root/data/disks/disk_s3_plain/test_table_backup/data/test_database/test_table/all_1_1_0/columns.txt HTTP/1.1" 204 0 http://172.16.8.4:9001 "DELETE /root/data/disks/disk_s3_plain/test_table_backup/data/test_database/test_table/all_1_1_0/count.txt HTTP/1.1" 204 0 http://172.16.8.4:9001 "DELETE /root/data/disks/disk_s3_plain/test_table_backup/data/test_database/test_table/all_1_1_0/data.bin HTTP/1.1" 204 0 http://172.16.8.4:9001 "DELETE /root/data/disks/disk_s3_plain/test_table_backup/data/test_database/test_table/all_1_1_0/data.cmrk3 HTTP/1.1" 204 0 http://172.16.8.4:9001 "DELETE /root/data/disks/disk_s3_plain/test_table_backup/data/test_database/test_table/all_1_1_0/default_compression_codec.txt HTTP/1.1" 204 0 http://172.16.8.4:9001 "DELETE /root/data/disks/disk_s3_plain/test_table_backup/data/test_database/test_table/all_1_1_0/metadata_version.txt HTTP/1.1" 204 0 http://172.16.8.4:9001 "DELETE /root/data/disks/disk_s3_plain/test_table_backup/data/test_database/test_table/all_1_1_0/primary.cidx HTTP/1.1" 204 0 http://172.16.8.4:9001 "DELETE /root/data/disks/disk_s3_plain/test_table_backup/data/test_database/test_table/all_1_1_0/serialization.json HTTP/1.1" 204 0 http://172.16.8.4:9001 "DELETE /root/data/disks/disk_s3_plain/test_table_backup/metadata/test_database/test_table.sql HTTP/1.1" 204 0 [gw1] PASSED test_database_backup/test.py::test_database_backup_table[Disk('backup_disk_s3_plain', 'test_table_backup')] test_database_backup/test.py::test_database_backup_table[File('test_table_backup_file')] run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', 'rm -rf /backups/'] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c rm -rf /backups/] Executing query rename table system.query_log to system.query_log_29_test on node2 run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', 'rm -rf /local_plain/'] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c rm -rf /local_plain/] Executing query select dictHas('MySQL_complex_key_direct_', tuple(toUInt64(1),toString('world'))) on node Executing query DROP USER IF EXISTS u1, u2 ON CLUSTER 'cluster3' on node1 Executing query select dictGet('RedisHash_complex_key_direct_Float32_', 'Float32_', tuple(toUInt64(1),toString('world'))) on node http://172.16.8.4:9001 "GET /root?delimiter=&encoding-type=url&list-type=2&max-keys=1000&prefix= HTTP/1.1" 200 0 Executing query DROP DATABASE IF EXISTS test_database; DROP DATABASE IF EXISTS test_table_backup; CREATE DATABASE test_database; CREATE TABLE test_database.test_table (id UInt64, value String) ENGINE=MergeTree ORDER BY id; INSERT INTO test_database.test_table VALUES (0, 'test_database.test_table'); BACKUP TABLE test_database.test_table TO File('test_table_backup_file'); CREATE DATABASE test_table_backup ENGINE = Backup('test_database', File('test_table_backup_file')); on instance Executing query select dictGetUInt64OrDefault('LocalClickHouse_complex_key_direct_', 'UInt64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt64(88)) on local_node Executing query system flush logs on node2 Executing query select dictGetOrDefault('MySQL_complex_key_direct_', 'DateTime_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDateTime('2018-12-30 00:00:00')) on node Executing query select dictGetFloat32('RedisHash_complex_key_direct_Float32_', 'Float32_', tuple(toUInt64(1),toString('world'))) on node test_backup_restore_on_cluster/test.py::test_replicated_database Executing query CREATE DATABASE mydb ON CLUSTER 'cluster' ENGINE=Replicated('/clickhouse/path/','{shard}','{replica}') on node1 Executing query select dictGet('LocalClickHouse_complex_key_direct_', 'Int8_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query SELECT id, value FROM test_table_backup.test_table on instance Executing query select dictGetDateTimeOrDefault('MySQL_complex_key_direct_', 'DateTime_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDateTime('2018-12-30 00:00:00')) on node Executing query select dictHas('RedisHash_complex_key_direct_Float32_', tuple(toUInt64(1),toString('world'))) on node run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:True cmd: ['bash', '-c', 'ps -C clickhouse'] Command:[docker exec -u root roottestdatabasebackup-gw1-instance-1 bash -c ps -C clickhouse] Executing query select dictGetInt8('LocalClickHouse_complex_key_direct_', 'Int8_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Stdout: PID TTY TIME CMD Stdout: 5784 ? 00:00:06 clickhouse run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', 'pkill clickhouse'] Command:[docker exec -u root roottestdatabasebackup-gw1-instance-1 bash -c pkill clickhouse] Executing query CREATE TABLE mydb.tbl(x UInt8, y String) ENGINE=ReplicatedMergeTree ORDER BY x on node1 Executing query select dictGet('MySQL_complex_key_direct_', 'String_', tuple(toUInt64(1),toString('world'))) on node run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select dictGetOrDefault('RedisHash_complex_key_direct_Float32_', 'Float32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat32(555.11)) on node Stdout:5784 Executing query select dictHas('LocalClickHouse_complex_key_direct_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select dictGetString('MySQL_complex_key_direct_', 'String_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGetFloat32OrDefault('RedisHash_complex_key_direct_Float32_', 'Float32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat32(555.11)) on node Executing query select count() from system.tables where database = 'system' and name in ['query_log', 'text_log'] on node2 Executing query SYSTEM SYNC DATABASE REPLICA mydb on node2 Executing query select dictHas('MySQL_complex_key_direct_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGetOrDefault('LocalClickHouse_complex_key_direct_', 'Int8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt8(-55)) on local_node Executing query select dictGet('RedisHash_complex_key_direct_Float32_', 'Float32_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query INSERT INTO mydb.tbl VALUES (1, 'Don''t') on node1 Executing query rename table system.text_log to system.text_log_30_test on node2 Executing query select dictGetOrDefault('MySQL_complex_key_direct_', 'String_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toString('hi')) on node Executing query select dictGetInt8OrDefault('LocalClickHouse_complex_key_direct_', 'Int8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt8(-55)) on local_node Executing query select dictGetFloat32('RedisHash_complex_key_direct_Float32_', 'Float32_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query INSERT INTO mydb.tbl VALUES (2, 'count') on node2 Executing query rename table system.query_log to system.query_log_30_test on node2 run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select dictGetStringOrDefault('MySQL_complex_key_direct_', 'String_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toString('hi')) on node Stdout:5784 Executing query select dictHas('RedisHash_complex_key_direct_Float32_', tuple(toUInt64(2),toString('qwerty2'))) on node run container_id:roottestasyncloaddatabases-gw7-node2-1 detach:False nothrow:True cmd: ['bash', '-c', 'ps -C clickhouse'] Command:[docker exec -u root roottestasyncloaddatabases-gw7-node2-1 bash -c ps -C clickhouse] Executing query select dictGet('LocalClickHouse_complex_key_direct_', 'Int16_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query INSERT INTO mydb.tbl VALUES (3, 'your') on node1 Stdout: PID TTY TIME CMD Stdout: 1607 ? 00:00:33 clickhouse run container_id:roottestasyncloaddatabases-gw7-node2-1 detach:False nothrow:False cmd: ['bash', '-c', 'pkill clickhouse'] Command:[docker exec -u root roottestasyncloaddatabases-gw7-node2-1 bash -c pkill clickhouse] run container_id:roottestasyncloaddatabases-gw7-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw7-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:1607 Executing query select dictGetOrDefault('RedisHash_complex_key_direct_Float32_', 'Float32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat32(555.11)) on node Executing query select dictGet('MySQL_complex_key_direct_', 'Float32_', tuple(toUInt64(1),toString('world'))) on node Executing query INSERT INTO mydb.tbl VALUES (4, 'chickens') on node2 Executing query select dictGetInt16('LocalClickHouse_complex_key_direct_', 'Int16_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select dictGetFloat32OrDefault('RedisHash_complex_key_direct_Float32_', 'Float32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat32(555.11)) on node Executing query select dictGetFloat32('MySQL_complex_key_direct_', 'Float32_', tuple(toUInt64(1),toString('world'))) on node Executing query OPTIMIZE TABLE mydb.tbl ON CLUSTER 'cluster' FINAL on node1 Executing query select dictHas('LocalClickHouse_complex_key_direct_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query system reload dictionaries on node Executing query select dictHas('MySQL_complex_key_direct_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGetOrDefault('LocalClickHouse_complex_key_direct_', 'Int16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt16(-66)) on local_node run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query SYSTEM SYNC REPLICA ON CLUSTER 'cluster' mydb.tbl on node1 Stdout:5784 [gw0] PASSED test_dictionaries_redis/test.py::test_redis_dictionaries[Float32_] test_dictionaries_redis/test.py::test_redis_dictionaries[Float64_] Run test with id: 13 Executing query system reload dictionaries on node run container_id:roottestasyncloaddatabases-gw7-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw7-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select dictGetOrDefault('MySQL_complex_key_direct_', 'Float32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat32(555.11)) on node Stdout:1607 Executing query select dictGetInt16OrDefault('LocalClickHouse_complex_key_direct_', 'Int16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt16(-66)) on local_node Executing query system reload dictionary RedisSimple_flat_Float64_ on node Executing query SELECT * FROM mydb.tbl ORDER BY x on node1 Executing query select dictGetFloat32OrDefault('MySQL_complex_key_direct_', 'Float32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat32(555.11)) on node Executing query select dictGet('LocalClickHouse_complex_key_direct_', 'Int32_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select dictGet('RedisSimple_flat_Float64_', 'Float64_', toUInt64(1)) on node Executing query SELECT * FROM mydb.tbl ORDER BY x on node2 Executing query select dictGet('MySQL_complex_key_direct_', 'Float64_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGetInt32('LocalClickHouse_complex_key_direct_', 'Int32_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query BACKUP DATABASE mydb ON CLUSTER 'cluster' TO Disk('backups', '12') SETTINGS replica_num=2 on node1 Executing query select dictGetFloat64('RedisSimple_flat_Float64_', 'Float64_', toUInt64(1)) on node Executing query select dictGetFloat64('MySQL_complex_key_direct_', 'Float64_', tuple(toUInt64(1),toString('world'))) on node run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:5784 Executing query select dictHas('RedisSimple_flat_Float64_', toUInt64(1)) on node Executing query select dictHas('LocalClickHouse_complex_key_direct_', tuple(toUInt64(2),toString('qwerty2'))) on local_node run container_id:roottestasyncloaddatabases-gw7-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw7-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:1607 Executing query select dictHas('MySQL_complex_key_direct_', tuple(toUInt64(1),toString('world'))) on node Executing query DROP DATABASE mydb ON CLUSTER 'cluster' SYNC on node1 Executing query select dictGetOrDefault('RedisSimple_flat_Float64_', 'Float64_', toUInt64(9999999), toFloat64(777.11)) on node Executing query select dictGetOrDefault('LocalClickHouse_complex_key_direct_', 'Int32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt32(-77)) on local_node Executing query select dictGetOrDefault('MySQL_complex_key_direct_', 'Float64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat64(777.11)) on node Executing query select dictGetFloat64OrDefault('RedisSimple_flat_Float64_', 'Float64_', toUInt64(9999999), toFloat64(777.11)) on node Executing query RESTORE DATABASE mydb ON CLUSTER 'cluster' FROM Disk('backups', '12') on node1 Executing query select dictGetInt32OrDefault('LocalClickHouse_complex_key_direct_', 'Int32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt32(-77)) on local_node Executing query select dictGetFloat64OrDefault('MySQL_complex_key_direct_', 'Float64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat64(777.11)) on node Executing query select dictGet('RedisSimple_flat_Float64_', 'Float64_', toUInt64(2)) on node run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select dictGet('LocalClickHouse_complex_key_direct_', 'Int64_', tuple(toUInt64(2),toString('qwerty2'))) on local_node No clickhouse process running. Start new one. http://localhost:None "POST /v1.46/containers/roottestdatabasebackup-gw1-instance-1/exec HTTP/1.1" 201 74 http://localhost:None "POST /v1.46/exec/986eef4c45cd86b88e7096d5d9037ab0dc8a840b368c140f7d116cf3be06e7b8/start HTTP/1.1" 200 0 http://localhost:None "GET /v1.46/exec/986eef4c45cd86b88e7096d5d9037ab0dc8a840b368c140f7d116cf3be06e7b8/json HTTP/1.1" 200 586 Executing query select dictGet('MySQL_complex_key_direct_', 'UInt8_', tuple(toUInt64(2),toString('qwerty2'))) on node run container_id:roottestasyncloaddatabases-gw7-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw7-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:1607 Executing query select dictGetFloat64('RedisSimple_flat_Float64_', 'Float64_', toUInt64(2)) on node Executing query select dictGetInt64('LocalClickHouse_complex_key_direct_', 'Int64_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select dictGetUInt8('MySQL_complex_key_direct_', 'UInt8_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictHas('RedisSimple_flat_Float64_', toUInt64(2)) on node Executing query SYSTEM SYNC REPLICA ON CLUSTER 'cluster' mydb.tbl on node1 Executing query select dictHas('LocalClickHouse_complex_key_direct_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select dictHas('MySQL_complex_key_direct_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGetOrDefault('RedisSimple_flat_Float64_', 'Float64_', toUInt64(9999999), toFloat64(777.11)) on node Executing query select dictGetOrDefault('MySQL_complex_key_direct_', 'UInt8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt8(55)) on node Executing query select dictGetOrDefault('LocalClickHouse_complex_key_direct_', 'Int64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt64(-88)) on local_node Executing query SELECT * FROM mydb.tbl ORDER BY x on node1 Executing query select dictGetFloat64OrDefault('RedisSimple_flat_Float64_', 'Float64_', toUInt64(9999999), toFloat64(777.11)) on node Stderr: node1 Skipped - Image is already being pulled by node2 Stderr: node2 Pulling Stderr: node2 Pulled ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_aggregation_memory_efficient/_instances-0-gw4/.env --project-name roottestaggregationmemoryefficient-gw4 --file /ClickHouse/tests/integration/test_aggregation_memory_efficient/_instances-0-gw4/node1/docker-compose.yml --file /ClickHouse/tests/integration/test_aggregation_memory_efficient/_instances-0-gw4/node2/docker-compose.yml up -d --no-recreate') Command:[docker compose --env-file /ClickHouse/tests/integration/test_aggregation_memory_efficient/_instances-0-gw4/.env --project-name roottestaggregationmemoryefficient-gw4 --file /ClickHouse/tests/integration/test_aggregation_memory_efficient/_instances-0-gw4/node1/docker-compose.yml --file /ClickHouse/tests/integration/test_aggregation_memory_efficient/_instances-0-gw4/node2/docker-compose.yml up -d --no-recreate] run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:6613 Clickhouse process running. run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:6613 Executing query select 20 on instance run container_id:roottestasyncloaddatabases-gw7-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw7-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select dictGetUInt8OrDefault('MySQL_complex_key_direct_', 'UInt8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt8(55)) on node Stdout:1607 Executing query SELECT * FROM mydb.tbl ORDER BY x on node2 Executing query system reload dictionary RedisSimple_hashed_Float64_ on node Executing query select dictGetInt64OrDefault('LocalClickHouse_complex_key_direct_', 'Int64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt64(-88)) on local_node Executing query select dictGet('RedisSimple_hashed_Float64_', 'Float64_', toUInt64(1)) on node Executing query DROP TABLE IF EXISTS tbl ON CLUSTER 'cluster3' SYNC on node1 [gw2] PASSED test_backup_restore_on_cluster/test.py::test_replicated_database Executing query select dictGet('MySQL_complex_key_direct_', 'UInt16_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGet('LocalClickHouse_complex_key_direct_', 'UUID_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Stderr: Network roottestaggregationmemoryefficient-gw4_default Creating Stderr: Network roottestaggregationmemoryefficient-gw4_default Created Stderr: Container roottestaggregationmemoryefficient-gw4-node2-1 Creating Stderr: Container roottestaggregationmemoryefficient-gw4-node1-1 Creating Stderr: Container roottestaggregationmemoryefficient-gw4-node1-1 Created Stderr: Container roottestaggregationmemoryefficient-gw4-node2-1 Created Stderr: Container roottestaggregationmemoryefficient-gw4-node1-1 Starting Stderr: Container roottestaggregationmemoryefficient-gw4-node2-1 Starting Stderr: Container roottestaggregationmemoryefficient-gw4-node2-1 Started Stderr: Container roottestaggregationmemoryefficient-gw4-node1-1 Started ClickHouse instance created get_instance_ip instance_name=node1 http://localhost:None "GET /v1.46/containers/roottestaggregationmemoryefficient-gw4-node1-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node1 http://localhost:None "GET /v1.46/containers/roottestaggregationmemoryefficient-gw4-node1-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node1, ip: 172.16.1.2... http://localhost:None "GET /v1.46/containers/roottestaggregationmemoryefficient-gw4-node1-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/d14a47efb990425323cc33c7beb85c921d32163e8b2b31a10e075affa60f421e/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/d14a47efb990425323cc33c7beb85c921d32163e8b2b31a10e075affa60f421e/json HTTP/1.1" 200 None Executing query select dictGetUInt16('MySQL_complex_key_direct_', 'UInt16_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGetFloat64('RedisSimple_hashed_Float64_', 'Float64_', toUInt64(1)) on node Executing query select 20 on instance Executing query select dictGetUUID('LocalClickHouse_complex_key_direct_', 'UUID_', tuple(toUInt64(2),toString('qwerty2'))) on local_node http://localhost:None "GET /v1.46/containers/d14a47efb990425323cc33c7beb85c921d32163e8b2b31a10e075affa60f421e/json HTTP/1.1" 200 None Executing query DROP TABLE IF EXISTS tbl2 ON CLUSTER 'cluster3' SYNC on node1 http://localhost:None "GET /v1.46/containers/d14a47efb990425323cc33c7beb85c921d32163e8b2b31a10e075affa60f421e/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/d14a47efb990425323cc33c7beb85c921d32163e8b2b31a10e075affa60f421e/json HTTP/1.1" 200 None Executing query select dictHas('MySQL_complex_key_direct_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictHas('RedisSimple_hashed_Float64_', toUInt64(1)) on node Executing query select dictHas('LocalClickHouse_complex_key_direct_', tuple(toUInt64(2),toString('qwerty2'))) on local_node http://localhost:None "GET /v1.46/containers/d14a47efb990425323cc33c7beb85c921d32163e8b2b31a10e075affa60f421e/json HTTP/1.1" 200 None run container_id:roottestasyncloaddatabases-gw7-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw7-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query DROP DATABASE IF EXISTS mydb ON CLUSTER 'cluster3' SYNC on node1 run container_id:roottestasyncloaddatabases-gw7-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw7-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] http://localhost:None "GET /v1.46/containers/d14a47efb990425323cc33c7beb85c921d32163e8b2b31a10e075affa60f421e/json HTTP/1.1" 200 None No clickhouse process running. Start new one. http://localhost:None "POST /v1.46/containers/roottestasyncloaddatabases-gw7-node2-1/exec HTTP/1.1" 201 74 http://localhost:None "POST /v1.46/exec/db4673671728c2e9b0f294e2dad34176af6ca12635972a7a73f8ea868c7375c7/start HTTP/1.1" 200 0 http://localhost:None "GET /v1.46/exec/db4673671728c2e9b0f294e2dad34176af6ca12635972a7a73f8ea868c7375c7/json HTTP/1.1" 200 586 http://localhost:None "GET /v1.46/containers/d14a47efb990425323cc33c7beb85c921d32163e8b2b31a10e075affa60f421e/json HTTP/1.1" 200 None Executing query select dictGetOrDefault('RedisSimple_hashed_Float64_', 'Float64_', toUInt64(9999999), toFloat64(777.11)) on node Executing query select dictGetOrDefault('MySQL_complex_key_direct_', 'UInt16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt16(66)) on node http://localhost:None "GET /v1.46/containers/d14a47efb990425323cc33c7beb85c921d32163e8b2b31a10e075affa60f421e/json HTTP/1.1" 200 None Executing query select dictGetOrDefault('LocalClickHouse_complex_key_direct_', 'UUID_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUUID('550e8400-0000-0000-0000-000000000000')) on local_node Executing query select 20 on instance http://localhost:None "GET /v1.46/containers/d14a47efb990425323cc33c7beb85c921d32163e8b2b31a10e075affa60f421e/json HTTP/1.1" 200 None Executing query select dictGetFloat64OrDefault('RedisSimple_hashed_Float64_', 'Float64_', toUInt64(9999999), toFloat64(777.11)) on node Executing query DROP DATABASE IF EXISTS mydb2 ON CLUSTER 'cluster3' SYNC on node1 Executing query select dictGetUInt16OrDefault('MySQL_complex_key_direct_', 'UInt16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt16(66)) on node http://localhost:None "GET /v1.46/containers/d14a47efb990425323cc33c7beb85c921d32163e8b2b31a10e075affa60f421e/json HTTP/1.1" 200 None Executing query select dictGetUUIDOrDefault('LocalClickHouse_complex_key_direct_', 'UUID_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUUID('550e8400-0000-0000-0000-000000000000')) on local_node http://localhost:None "GET /v1.46/containers/d14a47efb990425323cc33c7beb85c921d32163e8b2b31a10e075affa60f421e/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/d14a47efb990425323cc33c7beb85c921d32163e8b2b31a10e075affa60f421e/json HTTP/1.1" 200 None Executing query select dictGet('RedisSimple_hashed_Float64_', 'Float64_', toUInt64(2)) on node Executing query select dictGet('MySQL_complex_key_direct_', 'UInt32_', tuple(toUInt64(2),toString('qwerty2'))) on node http://localhost:None "GET /v1.46/containers/d14a47efb990425323cc33c7beb85c921d32163e8b2b31a10e075affa60f421e/json HTTP/1.1" 200 None Executing query DROP USER IF EXISTS u1, u2 ON CLUSTER 'cluster3' on node1 Executing query select dictGet('LocalClickHouse_complex_key_direct_', 'Date_', tuple(toUInt64(2),toString('qwerty2'))) on local_node http://localhost:None "GET /v1.46/containers/d14a47efb990425323cc33c7beb85c921d32163e8b2b31a10e075affa60f421e/json HTTP/1.1" 200 None Executing query select 20 on instance http://localhost:None "GET /v1.46/containers/d14a47efb990425323cc33c7beb85c921d32163e8b2b31a10e075affa60f421e/json HTTP/1.1" 200 None Executing query select dictGetFloat64('RedisSimple_hashed_Float64_', 'Float64_', toUInt64(2)) on node http://localhost:None "GET /v1.46/containers/d14a47efb990425323cc33c7beb85c921d32163e8b2b31a10e075affa60f421e/json HTTP/1.1" 200 None Executing query select dictGetUInt32('MySQL_complex_key_direct_', 'UInt32_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGetDate('LocalClickHouse_complex_key_direct_', 'Date_', tuple(toUInt64(2),toString('qwerty2'))) on local_node run container_id:roottestasyncloaddatabases-gw7-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw7-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:2441 Clickhouse process running. run container_id:roottestasyncloaddatabases-gw7-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw7-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] test_backup_restore_on_cluster/test.py::test_replicated_database_async Executing query CREATE DATABASE mydb ON CLUSTER 'cluster' ENGINE=Replicated('/clickhouse/path/','{shard}','{replica}') on node1 http://localhost:None "GET /v1.46/containers/d14a47efb990425323cc33c7beb85c921d32163e8b2b31a10e075affa60f421e/json HTTP/1.1" 200 None Stdout:2441 Executing query select 20 on node2 Executing query select dictHas('RedisSimple_hashed_Float64_', toUInt64(2)) on node http://localhost:None "GET /v1.46/containers/d14a47efb990425323cc33c7beb85c921d32163e8b2b31a10e075affa60f421e/json HTTP/1.1" 200 None Executing query select dictHas('MySQL_complex_key_direct_', tuple(toUInt64(2),toString('qwerty2'))) on node http://localhost:None "GET /v1.46/containers/d14a47efb990425323cc33c7beb85c921d32163e8b2b31a10e075affa60f421e/json HTTP/1.1" 200 None Executing query select dictHas('LocalClickHouse_complex_key_direct_', tuple(toUInt64(2),toString('qwerty2'))) on local_node http://localhost:None "GET /v1.46/containers/d14a47efb990425323cc33c7beb85c921d32163e8b2b31a10e075affa60f421e/json HTTP/1.1" 200 None Executing query CREATE TABLE mydb.tbl(x UInt8) ENGINE=ReplicatedMergeTree ORDER BY x on node1 Executing query select dictGetOrDefault('RedisSimple_hashed_Float64_', 'Float64_', toUInt64(9999999), toFloat64(777.11)) on node http://localhost:None "GET /v1.46/containers/d14a47efb990425323cc33c7beb85c921d32163e8b2b31a10e075affa60f421e/json HTTP/1.1" 200 None ClickHouse node1 started get_instance_ip instance_name=node2 http://localhost:None "GET /v1.46/containers/roottestaggregationmemoryefficient-gw4-node2-1/json HTTP/1.1" 200 None Executing query select 20 on instance get_instance_ip instance_name=node2 http://localhost:None "GET /v1.46/containers/roottestaggregationmemoryefficient-gw4-node2-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node2, ip: 172.16.1.3... http://localhost:None "GET /v1.46/containers/roottestaggregationmemoryefficient-gw4-node2-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/c56369e7205f354c7a658dfa08622d480976f9f179596cddcdd06bab6ac7f6a7/json HTTP/1.1" 200 None Executing query select dictGetOrDefault('MySQL_complex_key_direct_', 'UInt32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt32(77)) on node http://localhost:None "GET /v1.46/containers/c56369e7205f354c7a658dfa08622d480976f9f179596cddcdd06bab6ac7f6a7/json HTTP/1.1" 200 None ClickHouse node2 started Executing query create table da_memory_efficient_shard(A Int64, B Int64) Engine=MergeTree order by A partition by B % 2; on node1 Executing query select dictGetOrDefault('LocalClickHouse_complex_key_direct_', 'Date_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDate('2018-12-30')) on local_node Executing query select dictGetFloat64OrDefault('RedisSimple_hashed_Float64_', 'Float64_', toUInt64(9999999), toFloat64(777.11)) on node Executing query SELECT id, value FROM test_table_backup.test_table on instance Executing query select dictGetUInt32OrDefault('MySQL_complex_key_direct_', 'UInt32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt32(77)) on node Executing query select 20 on node2 Executing query create table da_memory_efficient_shard(A Int64, B Int64) Engine=MergeTree order by A partition by B % 2; on node2 Executing query select dictGetDateOrDefault('LocalClickHouse_complex_key_direct_', 'Date_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDate('2018-12-30')) on local_node Executing query system reload dictionary RedisSimple_cache_Float64_ on node Executing query insert into da_memory_efficient_shard select number, number from numbers(100000); on node1 Executing query DROP DATABASE test_table_backup on instance Executing query select dictGet('MySQL_complex_key_direct_', 'UInt64_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGet('RedisSimple_cache_Float64_', 'Float64_', toUInt64(1)) on node Executing query select dictGet('LocalClickHouse_complex_key_direct_', 'DateTime_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query DROP DATABASE test_database on instance Executing query CREATE TABLE mydb.tbl2(y String) ENGINE=ReplicatedMergeTree ORDER BY y on node1 Executing query select dictGetUInt64('MySQL_complex_key_direct_', 'UInt64_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query insert into da_memory_efficient_shard select number + 100000, number from numbers(100000); on node2 Executing query select dictGetFloat64('RedisSimple_cache_Float64_', 'Float64_', toUInt64(1)) on node run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', 'rm -rf /backups/'] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c rm -rf /backups/] Executing query select 20 on node2 run container_id:roottestdatabasebackup-gw1-instance-1 detach:False nothrow:False cmd: ['bash', '-c', 'rm -rf /local_plain/'] Command:[docker exec roottestdatabasebackup-gw1-instance-1 bash -c rm -rf /local_plain/] Executing query select dictGetDateTime('LocalClickHouse_complex_key_direct_', 'DateTime_', tuple(toUInt64(2),toString('qwerty2'))) on local_node http://172.16.8.4:9001 "GET /root?delimiter=&encoding-type=url&list-type=2&max-keys=1000&prefix= HTTP/1.1" 200 0 Command:[docker compose --env-file /ClickHouse/tests/integration/test_database_backup/_instances-0-gw1/.env --project-name roottestdatabasebackup-gw1 --file /ClickHouse/tests/integration/test_database_backup/_instances-0-gw1/instance/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_minio.yml stop --timeout 20] [gw1] PASSED test_database_backup/test.py::test_database_backup_table[File('test_table_backup_file')] Executing query SYSTEM SYNC DATABASE REPLICA mydb on node2 Executing query select dictHas('MySQL_complex_key_direct_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictHas('RedisSimple_cache_Float64_', toUInt64(1)) on node Executing query set distributed_aggregation_memory_efficient = 1, group_by_two_level_threshold = 1, group_by_two_level_threshold_bytes=1 on node1 Executing query INSERT INTO mydb.tbl VALUES (1) on node1 Executing query select dictHas('LocalClickHouse_complex_key_direct_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select dictGetOrDefault('RedisSimple_cache_Float64_', 'Float64_', toUInt64(9999999), toFloat64(777.11)) on node Executing query select sum(a) from (SELECT B, uniqExact(A) a FROM remote('node{1,2}', default.da_memory_efficient_shard) GROUP BY B) on node1 Executing query select dictGetOrDefault('MySQL_complex_key_direct_', 'UInt64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt64(88)) on node Executing query INSERT INTO mydb.tbl VALUES (22) on node1 Executing query select 20 on node2 Executing query select dictGetOrDefault('LocalClickHouse_complex_key_direct_', 'DateTime_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDateTime('2018-12-30 00:00:00')) on local_node Executing query select dictGetFloat64OrDefault('RedisSimple_cache_Float64_', 'Float64_', toUInt64(9999999), toFloat64(777.11)) on node Executing query INSERT INTO mydb.tbl2 VALUES ('a') on node2 Executing query select dictGetUInt64OrDefault('MySQL_complex_key_direct_', 'UInt64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt64(88)) on node Executing query select dictGetDateTimeOrDefault('LocalClickHouse_complex_key_direct_', 'DateTime_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDateTime('2018-12-30 00:00:00')) on local_node Executing query INSERT INTO mydb.tbl2 VALUES ('bb') on node2 Executing query select dictGet('RedisSimple_cache_Float64_', 'Float64_', toUInt64(2)) on node Executing query select dictGet('MySQL_complex_key_direct_', 'Int8_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select 20 on node2 Executing query select dictGet('LocalClickHouse_complex_key_direct_', 'String_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query OPTIMIZE TABLE mydb.tbl ON CLUSTER 'cluster' FINAL on node1 Executing query select dictGetInt8('MySQL_complex_key_direct_', 'Int8_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGetFloat64('RedisSimple_cache_Float64_', 'Float64_', toUInt64(2)) on node Executing query set distributed_aggregation_memory_efficient = 0 on node1 Executing query select dictGetString('LocalClickHouse_complex_key_direct_', 'String_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select dictHas('RedisSimple_cache_Float64_', toUInt64(2)) on node Executing query select dictHas('MySQL_complex_key_direct_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select sum(a) from (SELECT B, uniqExact(A) a FROM remote('node{1,2}', default.da_memory_efficient_shard) GROUP BY B) on node1 Executing query OPTIMIZE TABLE mydb.tbl2 ON CLUSTER 'cluster' FINAL on node1 Executing query select dictHas('LocalClickHouse_complex_key_direct_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select dictGetOrDefault('RedisSimple_cache_Float64_', 'Float64_', toUInt64(9999999), toFloat64(777.11)) on node Executing query select dictGetOrDefault('MySQL_complex_key_direct_', 'Int8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt8(-55)) on node Executing query SYSTEM SYNC REPLICA ON CLUSTER 'cluster' mydb.tbl on node1 Executing query select dictGetOrDefault('LocalClickHouse_complex_key_direct_', 'String_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toString('hi')) on local_node Executing query select dictGetFloat64OrDefault('RedisSimple_cache_Float64_', 'Float64_', toUInt64(9999999), toFloat64(777.11)) on node Executing query select dictGetInt8OrDefault('MySQL_complex_key_direct_', 'Int8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt8(-55)) on node Executing query SYSTEM SYNC REPLICA ON CLUSTER 'cluster' mydb.tbl2 on node1 Executing query system reload dictionary RedisSimple_direct_Float64_ on node Executing query select dictGetStringOrDefault('LocalClickHouse_complex_key_direct_', 'String_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toString('hi')) on local_node Executing query select dictGet('MySQL_complex_key_direct_', 'Int16_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select count() from system.text_log_25_test on node2 Executing query select dictGet('RedisSimple_direct_Float64_', 'Float64_', toUInt64(1)) on node Executing query set distributed_aggregation_memory_efficient = 1, group_by_two_level_threshold = 1, group_by_two_level_threshold_bytes=1 on node1 Executing query select dictGet('LocalClickHouse_complex_key_direct_', 'Float32_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query BACKUP DATABASE mydb ON CLUSTER 'cluster' TO Disk('backups', '13') ASYNC on node1 Executing query select count() from system.query_log_7_test on node2 Executing query select dictGetInt16('MySQL_complex_key_direct_', 'Int16_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query SELECT fullHostName() AS h, uniqExact(A) AS a FROM remote('node{1,2}', default.da_memory_efficient_shard) GROUP BY h ORDER BY h; on node1 Executing query select dictGetFloat64('RedisSimple_direct_Float64_', 'Float64_', toUInt64(1)) on node Executing query SELECT status, error FROM system.backups WHERE id='1d147db6-c182-4255-a150-5d0fc28335a7' on node1 Executing query select dictHas('MySQL_complex_key_direct_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGetFloat32('LocalClickHouse_complex_key_direct_', 'Float32_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select count() from system.text_log_15_test on node2 Executing query select dictHas('RedisSimple_direct_Float64_', toUInt64(1)) on node Executing query set distributed_aggregation_memory_efficient = 0 on node1 Executing query select count() from system.query_log_19_test on node2 Executing query select dictGetOrDefault('MySQL_complex_key_direct_', 'Int16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt16(-66)) on node Executing query select dictHas('LocalClickHouse_complex_key_direct_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select dictGetOrDefault('RedisSimple_direct_Float64_', 'Float64_', toUInt64(9999999), toFloat64(777.11)) on node Executing query SELECT fullHostName() AS h, uniqExact(A) AS a FROM remote('node{1,2}', default.da_memory_efficient_shard) GROUP BY h ORDER BY h; on node1 Executing query select dictGetInt16OrDefault('MySQL_complex_key_direct_', 'Int16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt16(-66)) on node Executing query select count() from system.text_log_18_test on node2 Executing query select dictGetOrDefault('LocalClickHouse_complex_key_direct_', 'Float32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat32(555.11)) on local_node Executing query SELECT status, error FROM system.backups WHERE id='1d147db6-c182-4255-a150-5d0fc28335a7' on node1 Executing query select dictGetFloat64OrDefault('RedisSimple_direct_Float64_', 'Float64_', toUInt64(9999999), toFloat64(777.11)) on node Executing query select dictGet('MySQL_complex_key_direct_', 'Int32_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select count() from system.query_log_4_test on node2 Executing query select dictGetFloat32OrDefault('LocalClickHouse_complex_key_direct_', 'Float32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat32(555.11)) on local_node Command:[docker compose --env-file /ClickHouse/tests/integration/test_aggregation_memory_efficient/_instances-0-gw4/.env --project-name roottestaggregationmemoryefficient-gw4 --file /ClickHouse/tests/integration/test_aggregation_memory_efficient/_instances-0-gw4/node1/docker-compose.yml --file /ClickHouse/tests/integration/test_aggregation_memory_efficient/_instances-0-gw4/node2/docker-compose.yml stop --timeout 20] [gw4] PASSED test_aggregation_memory_efficient/test.py::test_remote Executing query DROP DATABASE mydb ON CLUSTER 'cluster' SYNC on node1 Executing query select dictGet('RedisSimple_direct_Float64_', 'Float64_', toUInt64(2)) on node Executing query select dictGetInt32('MySQL_complex_key_direct_', 'Int32_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select count() from system.asynchronous_loader where job ilike '%_log_%_test' and execution_pool = 'BackgroundLoad' on node2 Executing query select dictGet('LocalClickHouse_complex_key_direct_', 'Float64_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select dictGetFloat64('RedisSimple_direct_Float64_', 'Float64_', toUInt64(2)) on node Executing query select dictHas('MySQL_complex_key_direct_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query system flush logs on node2 Executing query RESTORE DATABASE mydb ON CLUSTER 'cluster' FROM Disk('backups', '13') ASYNC on node1 Executing query select dictGetFloat64('LocalClickHouse_complex_key_direct_', 'Float64_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select dictHas('RedisSimple_direct_Float64_', toUInt64(2)) on node Executing query select dictGetOrDefault('MySQL_complex_key_direct_', 'Int32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt32(-77)) on node Executing query SELECT status, error FROM system.backups WHERE id='7c297d83-40fc-41f0-972c-cf27e142b1d4' on node1 Executing query select dictGetOrDefault('RedisSimple_direct_Float64_', 'Float64_', toUInt64(9999999), toFloat64(777.11)) on node Executing query select dictHas('LocalClickHouse_complex_key_direct_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select dictGetInt32OrDefault('MySQL_complex_key_direct_', 'Int32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt32(-77)) on node Executing query select dictGetFloat64OrDefault('RedisSimple_direct_Float64_', 'Float64_', toUInt64(9999999), toFloat64(777.11)) on node Executing query select dictGetOrDefault('LocalClickHouse_complex_key_direct_', 'Float64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat64(777.11)) on local_node Executing query select dictGet('MySQL_complex_key_direct_', 'Int64_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select count() from system.tables where database = 'system' and name in ['query_log', 'text_log'] on node2 Executing query system reload dictionary RedisHash_complex_key_hashed_Float64_ on node Executing query select dictGetFloat64OrDefault('LocalClickHouse_complex_key_direct_', 'Float64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat64(777.11)) on local_node Executing query select dictGetInt64('MySQL_complex_key_direct_', 'Int64_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query SELECT status, error FROM system.backups WHERE id='7c297d83-40fc-41f0-972c-cf27e142b1d4' on node1 Executing query select dictGet('RedisHash_complex_key_hashed_Float64_', 'Float64_', tuple(toUInt64(1),toString('world'))) on node Executing query rename table system.text_log to system.text_log_31_test on node2 [gw5] PASSED test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_complex[complex_key_direct] test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_complex[complex_key_hashed] Executing query system reload dictionaries on local_node Executing query select dictHas('MySQL_complex_key_direct_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query rename table system.query_log to system.query_log_31_test on node2 Executing query select dictGetFloat64('RedisHash_complex_key_hashed_Float64_', 'Float64_', tuple(toUInt64(1),toString('world'))) on node Executing query SYSTEM SYNC REPLICA ON CLUSTER 'cluster' mydb.tbl on node1 Executing query select dictGet('LocalClickHouse_complex_key_hashed_', 'UInt8_', tuple(toUInt64(1),toString('world'))) on local_node Executing query system flush logs on node2 Stderr: Container roottestaggregationmemoryefficient-gw4-node2-1 Stopping Executing query select dictGetOrDefault('MySQL_complex_key_direct_', 'Int64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt64(-88)) on node Stderr: Container roottestaggregationmemoryefficient-gw4-node1-1 Stopping Stderr: Container roottestaggregationmemoryefficient-gw4-node1-1 Stopped Stderr: Container roottestaggregationmemoryefficient-gw4-node2-1 Stopped Command:[bash -c [ -f /ClickHouse/tests/integration/test_aggregation_memory_efficient/_instances-0-gw4/node1/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_aggregation_memory_efficient/_instances-0-gw4/node1/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[bash -c [ -f /ClickHouse/tests/integration/test_aggregation_memory_efficient/_instances-0-gw4/node2/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_aggregation_memory_efficient/_instances-0-gw4/node2/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[docker compose --env-file /ClickHouse/tests/integration/test_aggregation_memory_efficient/_instances-0-gw4/.env --project-name roottestaggregationmemoryefficient-gw4 --file /ClickHouse/tests/integration/test_aggregation_memory_efficient/_instances-0-gw4/node1/docker-compose.yml --file /ClickHouse/tests/integration/test_aggregation_memory_efficient/_instances-0-gw4/node2/docker-compose.yml down --volumes] Executing query select dictHas('RedisHash_complex_key_hashed_Float64_', tuple(toUInt64(1),toString('world'))) on node Executing query SYSTEM SYNC REPLICA ON CLUSTER 'cluster' mydb.tbl2 on node1 Executing query select dictGetUInt8('LocalClickHouse_complex_key_hashed_', 'UInt8_', tuple(toUInt64(1),toString('world'))) on local_node Executing query select dictGetInt64OrDefault('MySQL_complex_key_direct_', 'Int64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt64(-88)) on node Executing query select dictGetOrDefault('RedisHash_complex_key_hashed_Float64_', 'Float64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat64(777.11)) on node Executing query SELECT * FROM mydb.tbl ORDER BY x on node1 Executing query select dictHas('LocalClickHouse_complex_key_hashed_', tuple(toUInt64(1),toString('world'))) on local_node Executing query select dictGet('MySQL_complex_key_direct_', 'UUID_', tuple(toUInt64(2),toString('qwerty2'))) on node Stderr: Container roottestaggregationmemoryefficient-gw4-node1-1 Stopping Stderr: Container roottestaggregationmemoryefficient-gw4-node2-1 Stopping Stderr: Container roottestaggregationmemoryefficient-gw4-node1-1 Stopped Stderr: Container roottestaggregationmemoryefficient-gw4-node1-1 Removing Stderr: Container roottestaggregationmemoryefficient-gw4-node2-1 Stopped Stderr: Container roottestaggregationmemoryefficient-gw4-node2-1 Removing Stderr: Container roottestaggregationmemoryefficient-gw4-node2-1 Removed Stderr: Container roottestaggregationmemoryefficient-gw4-node1-1 Removed Stderr: Network roottestaggregationmemoryefficient-gw4_default Removing Stderr: Network roottestaggregationmemoryefficient-gw4_default Removed Cleanup called Docker networks for project roottestaggregationmemoryefficient-gw4 are NETWORK ID NAME DRIVER SCOPE Executing query select dictGetFloat64OrDefault('RedisHash_complex_key_hashed_Float64_', 'Float64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat64(777.11)) on node Docker containers for project roottestaggregationmemoryefficient-gw4 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestaggregationmemoryefficient-gw4 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestaggregationmemoryefficient-gw4-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} No running containers for project: roottestaggregationmemoryefficient-gw4 Trying to prune unused networks... Trying to prune unused images... Command:[docker image prune -f] Executing query SELECT * FROM mydb.tbl2 ORDER BY y on node2 Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Stdout:5 Command:[docker volume prune -f] Stdout:Total reclaimed space: 0B Volumes pruned: 5 test_analyzer_compatibility/test.py::test_two_new_versions Running tests in /ClickHouse/tests/integration/test_analyzer_compatibility/test.py Cluster start called. is_up=False Docker networks for project roottestanalyzercompatibility-gw4 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestanalyzercompatibility-gw4 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestanalyzercompatibility-gw4 are DRIVER VOLUME NAME Cleanup called Docker networks for project roottestanalyzercompatibility-gw4 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestanalyzercompatibility-gw4 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestanalyzercompatibility-gw4 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestanalyzercompatibility-gw4-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} No running containers for project: roottestanalyzercompatibility-gw4 Trying to prune unused networks... Executing query select dictGetUUID('MySQL_complex_key_direct_', 'UUID_', tuple(toUInt64(2),toString('qwerty2'))) on node Trying to prune unused images... Command:[docker image prune -f] Executing query select count() from system.tables where database = 'system' and name in ['query_log', 'text_log'] on node2 Executing query select dictGetOrDefault('LocalClickHouse_complex_key_hashed_', 'UInt8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt8(55)) on local_node Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Stdout:5 Command:[docker volume prune -f] Stdout:Total reclaimed space: 0B Volumes pruned: 5 Setup directory for instance: current Create directory for configuration generated in this helper Create directory for common tests configuration Copy common configuration from helpers Generate and write macros file Copy custom test config files ['/ClickHouse/tests/integration/test_analyzer_compatibility/configs/remote_servers.xml'] to /ClickHouse/tests/integration/test_analyzer_compatibility/_instances-0-gw4/current/configs/config.d Setup database dir /ClickHouse/tests/integration/test_analyzer_compatibility/_instances-0-gw4/current/database Setup logs dir /ClickHouse/tests/integration/test_analyzer_compatibility/_instances-0-gw4/current/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Setup directory for instance: backward Create directory for configuration generated in this helper Create directory for common tests configuration Copy common configuration from helpers Generate and write macros file Copy custom test config files ['/ClickHouse/tests/integration/test_analyzer_compatibility/configs/remote_servers.xml'] to /ClickHouse/tests/integration/test_analyzer_compatibility/_instances-0-gw4/backward/configs/config.d Setup database dir /ClickHouse/tests/integration/test_analyzer_compatibility/_instances-0-gw4/backward/database Setup logs dir /ClickHouse/tests/integration/test_analyzer_compatibility/_instances-0-gw4/backward/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw'} stored in /ClickHouse/tests/integration/test_analyzer_compatibility/_instances-0-gw4/.env Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found Executing query select dictGet('RedisHash_complex_key_hashed_Float64_', 'Float64_', tuple(toUInt64(2),toString('qwerty2'))) on node http://localhost:None "GET /version HTTP/1.1" 200 826 Command:[docker compose --env-file /ClickHouse/tests/integration/test_analyzer_compatibility/_instances-0-gw4/.env --project-name roottestanalyzercompatibility-gw4 --file /ClickHouse/tests/integration/test_analyzer_compatibility/_instances-0-gw4/current/docker-compose.yml --file /ClickHouse/tests/integration/test_analyzer_compatibility/_instances-0-gw4/backward/docker-compose.yml pull] Executing query DROP TABLE IF EXISTS tbl ON CLUSTER 'cluster3' SYNC on node1 [gw2] PASSED test_backup_restore_on_cluster/test.py::test_replicated_database_async Executing query select dictHas('MySQL_complex_key_direct_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query rename table system.text_log to system.text_log_32_test on node2 Executing query select dictGetUInt8OrDefault('LocalClickHouse_complex_key_hashed_', 'UInt8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt8(55)) on local_node Executing query select dictGetFloat64('RedisHash_complex_key_hashed_Float64_', 'Float64_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query DROP TABLE IF EXISTS tbl2 ON CLUSTER 'cluster3' SYNC on node1 Executing query rename table system.query_log to system.query_log_32_test on node2 Executing query select dictGet('LocalClickHouse_complex_key_hashed_', 'UInt16_', tuple(toUInt64(1),toString('world'))) on local_node Executing query select dictHas('RedisHash_complex_key_hashed_Float64_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGetOrDefault('MySQL_complex_key_direct_', 'UUID_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUUID('550e8400-0000-0000-0000-000000000000')) on node Executing query system flush logs on node2 Executing query select dictGetUInt16('LocalClickHouse_complex_key_hashed_', 'UInt16_', tuple(toUInt64(1),toString('world'))) on local_node Executing query DROP DATABASE IF EXISTS mydb ON CLUSTER 'cluster3' SYNC on node1 Executing query select dictGetOrDefault('RedisHash_complex_key_hashed_Float64_', 'Float64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat64(777.11)) on node Executing query select dictGetUUIDOrDefault('MySQL_complex_key_direct_', 'UUID_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUUID('550e8400-0000-0000-0000-000000000000')) on node Executing query select dictHas('LocalClickHouse_complex_key_hashed_', tuple(toUInt64(1),toString('world'))) on local_node Executing query select dictGetFloat64OrDefault('RedisHash_complex_key_hashed_Float64_', 'Float64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat64(777.11)) on node Executing query select dictGet('MySQL_complex_key_direct_', 'Date_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query DROP DATABASE IF EXISTS mydb2 ON CLUSTER 'cluster3' SYNC on node1 Executing query select dictGetOrDefault('LocalClickHouse_complex_key_hashed_', 'UInt16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt16(66)) on local_node Executing query system reload dictionary RedisHash_complex_key_cache_Float64_ on node Executing query select dictGetDate('MySQL_complex_key_direct_', 'Date_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGetUInt16OrDefault('LocalClickHouse_complex_key_hashed_', 'UInt16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt16(66)) on local_node Executing query select dictGet('RedisHash_complex_key_cache_Float64_', 'Float64_', tuple(toUInt64(1),toString('world'))) on node Executing query DROP USER IF EXISTS u1, u2 ON CLUSTER 'cluster3' on node1 Executing query select count() from system.tables where database = 'system' and name in ['query_log', 'text_log'] on node2 Executing query select dictHas('MySQL_complex_key_direct_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGet('LocalClickHouse_complex_key_hashed_', 'UInt32_', tuple(toUInt64(1),toString('world'))) on local_node Executing query select dictGetFloat64('RedisHash_complex_key_cache_Float64_', 'Float64_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGetOrDefault('MySQL_complex_key_direct_', 'Date_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDate('2018-12-30')) on node test_backup_restore_on_cluster/test.py::test_replicated_database_compare_parts Executing query CREATE DATABASE mydb ON CLUSTER 'cluster' ENGINE=Replicated('/clickhouse/path/','{shard}','{replica}') on node1 Executing query rename table system.text_log to system.text_log_33_test on node2 Executing query select dictGetUInt32('LocalClickHouse_complex_key_hashed_', 'UInt32_', tuple(toUInt64(1),toString('world'))) on local_node Executing query select dictHas('RedisHash_complex_key_cache_Float64_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGetDateOrDefault('MySQL_complex_key_direct_', 'Date_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDate('2018-12-30')) on node Executing query rename table system.query_log to system.query_log_33_test on node2 Executing query select dictHas('LocalClickHouse_complex_key_hashed_', tuple(toUInt64(1),toString('world'))) on local_node Executing query select dictGetOrDefault('RedisHash_complex_key_cache_Float64_', 'Float64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat64(777.11)) on node Executing query CREATE TABLE mydb.tbl(x UInt8, y String) ENGINE=ReplicatedMergeTree ORDER BY x on node1 Executing query system flush logs on node2 Executing query select dictGet('MySQL_complex_key_direct_', 'DateTime_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGetOrDefault('LocalClickHouse_complex_key_hashed_', 'UInt32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt32(77)) on local_node Executing query select dictGetFloat64OrDefault('RedisHash_complex_key_cache_Float64_', 'Float64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat64(777.11)) on node Executing query select dictGetDateTime('MySQL_complex_key_direct_', 'DateTime_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGetUInt32OrDefault('LocalClickHouse_complex_key_hashed_', 'UInt32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt32(77)) on local_node Executing query SYSTEM SYNC DATABASE REPLICA mydb on node2 Executing query select dictGet('RedisHash_complex_key_cache_Float64_', 'Float64_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query SYSTEM STOP MERGES mydb.tbl on node1 Executing query select dictHas('MySQL_complex_key_direct_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGet('LocalClickHouse_complex_key_hashed_', 'UInt64_', tuple(toUInt64(1),toString('world'))) on local_node Executing query select dictGetFloat64('RedisHash_complex_key_cache_Float64_', 'Float64_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query SYSTEM STOP MERGES mydb.tbl on node2 Executing query select count() from system.tables where database = 'system' and name in ['query_log', 'text_log'] on node2 Executing query select dictGetUInt64('LocalClickHouse_complex_key_hashed_', 'UInt64_', tuple(toUInt64(1),toString('world'))) on local_node Executing query select dictGetOrDefault('MySQL_complex_key_direct_', 'DateTime_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDateTime('2018-12-30 00:00:00')) on node Executing query select dictHas('RedisHash_complex_key_cache_Float64_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query SYSTEM STOP FETCHES mydb.tbl on node1 Executing query select dictHas('LocalClickHouse_complex_key_hashed_', tuple(toUInt64(1),toString('world'))) on local_node Executing query rename table system.text_log to system.text_log_34_test on node2 Executing query select dictGetDateTimeOrDefault('MySQL_complex_key_direct_', 'DateTime_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDateTime('2018-12-30 00:00:00')) on node Executing query select dictGetOrDefault('RedisHash_complex_key_cache_Float64_', 'Float64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat64(777.11)) on node Executing query SYSTEM STOP FETCHES mydb.tbl on node2 Executing query select dictGetOrDefault('LocalClickHouse_complex_key_hashed_', 'UInt64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt64(88)) on local_node Executing query rename table system.query_log to system.query_log_34_test on node2 Executing query INSERT INTO mydb.tbl VALUES (1, 'a') on node1 Executing query select dictGet('MySQL_complex_key_direct_', 'String_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGetFloat64OrDefault('RedisHash_complex_key_cache_Float64_', 'Float64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat64(777.11)) on node Executing query select dictGetUInt64OrDefault('LocalClickHouse_complex_key_hashed_', 'UInt64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt64(88)) on local_node Executing query system flush logs on node2 Executing query INSERT INTO mydb.tbl VALUES (2, 'b') on node1 Executing query select dictGetString('MySQL_complex_key_direct_', 'String_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query system reload dictionary RedisHash_complex_key_direct_Float64_ on node Executing query select dictGet('LocalClickHouse_complex_key_hashed_', 'Int8_', tuple(toUInt64(1),toString('world'))) on local_node Executing query INSERT INTO mydb.tbl VALUES (3, 'x') on node2 Executing query select dictHas('MySQL_complex_key_direct_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGet('RedisHash_complex_key_direct_Float64_', 'Float64_', tuple(toUInt64(1),toString('world'))) on node Executing query INSERT INTO mydb.tbl VALUES (4, 'y') on node2 Executing query select dictGetInt8('LocalClickHouse_complex_key_hashed_', 'Int8_', tuple(toUInt64(1),toString('world'))) on local_node Executing query select dictGetOrDefault('MySQL_complex_key_direct_', 'String_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toString('hi')) on node Executing query select dictGetFloat64('RedisHash_complex_key_direct_Float64_', 'Float64_', tuple(toUInt64(1),toString('world'))) on node Executing query SELECT * FROM mydb.tbl ORDER BY x on node2 Executing query select dictHas('LocalClickHouse_complex_key_hashed_', tuple(toUInt64(1),toString('world'))) on local_node Executing query select dictGetStringOrDefault('MySQL_complex_key_direct_', 'String_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toString('hi')) on node Executing query select dictHas('RedisHash_complex_key_direct_Float64_', tuple(toUInt64(1),toString('world'))) on node Executing query select count() from system.tables where database = 'system' and name in ['query_log', 'text_log'] on node2 Executing query BACKUP DATABASE mydb ON CLUSTER 'cluster' TO Disk('backups', '14') SETTINGS replica_num=2 on node1 Executing query select dictGetOrDefault('LocalClickHouse_complex_key_hashed_', 'Int8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt8(-55)) on local_node Executing query select dictGet('MySQL_complex_key_direct_', 'Float32_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGetOrDefault('RedisHash_complex_key_direct_Float64_', 'Float64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat64(777.11)) on node Executing query rename table system.text_log to system.text_log_35_test on node2 Executing query select dictGetInt8OrDefault('LocalClickHouse_complex_key_hashed_', 'Int8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt8(-55)) on local_node Executing query select dictGetFloat32('MySQL_complex_key_direct_', 'Float32_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGetFloat64OrDefault('RedisHash_complex_key_direct_Float64_', 'Float64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat64(777.11)) on node Executing query rename table system.query_log to system.query_log_35_test on node2 Executing query select dictGet('LocalClickHouse_complex_key_hashed_', 'Int16_', tuple(toUInt64(1),toString('world'))) on local_node Executing query DROP DATABASE mydb ON CLUSTER 'cluster' SYNC on node1 Executing query select dictHas('MySQL_complex_key_direct_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query system flush logs on node2 Executing query select dictGet('RedisHash_complex_key_direct_Float64_', 'Float64_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGetInt16('LocalClickHouse_complex_key_hashed_', 'Int16_', tuple(toUInt64(1),toString('world'))) on local_node Executing query select dictGetOrDefault('MySQL_complex_key_direct_', 'Float32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat32(555.11)) on node Executing query RESTORE DATABASE mydb ON CLUSTER 'cluster' FROM Disk('backups', '14') SETTINGS replica_num_in_backup=2 on node1 Executing query select dictGetFloat64('RedisHash_complex_key_direct_Float64_', 'Float64_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictHas('LocalClickHouse_complex_key_hashed_', tuple(toUInt64(1),toString('world'))) on local_node Executing query select dictGetFloat32OrDefault('MySQL_complex_key_direct_', 'Float32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat32(555.11)) on node Executing query select dictHas('RedisHash_complex_key_direct_Float64_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGetOrDefault('LocalClickHouse_complex_key_hashed_', 'Int16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt16(-66)) on local_node Executing query select dictGet('MySQL_complex_key_direct_', 'Float64_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select count() from system.tables where database = 'system' and name in ['query_log', 'text_log'] on node2 Executing query select dictGetOrDefault('RedisHash_complex_key_direct_Float64_', 'Float64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat64(777.11)) on node Executing query select dictGetInt16OrDefault('LocalClickHouse_complex_key_hashed_', 'Int16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt16(-66)) on local_node Executing query select dictGetFloat64('MySQL_complex_key_direct_', 'Float64_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query rename table system.text_log to system.text_log_36_test on node2 Executing query select dictGetFloat64OrDefault('RedisHash_complex_key_direct_Float64_', 'Float64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat64(777.11)) on node Executing query SYSTEM SYNC REPLICA ON CLUSTER 'cluster' mydb.tbl on node1 Executing query select dictGet('LocalClickHouse_complex_key_hashed_', 'Int32_', tuple(toUInt64(1),toString('world'))) on local_node Executing query select dictHas('MySQL_complex_key_direct_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query rename table system.query_log to system.query_log_36_test on node2 Executing query system reload dictionaries on node Executing query select dictGetOrDefault('MySQL_complex_key_direct_', 'Float64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat64(777.11)) on node Executing query SELECT _part, * FROM mydb.tbl ORDER BY x on node1 Executing query select dictGetInt32('LocalClickHouse_complex_key_hashed_', 'Int32_', tuple(toUInt64(1),toString('world'))) on local_node Executing query system flush logs on node2 [gw0] PASSED test_dictionaries_redis/test.py::test_redis_dictionaries[Float64_] test_dictionaries_redis/test.py::test_redis_dictionaries[Int16_] Run test with id: 5 Executing query system reload dictionaries on node Executing query select dictGetFloat64OrDefault('MySQL_complex_key_direct_', 'Float64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat64(777.11)) on node Executing query select dictHas('LocalClickHouse_complex_key_hashed_', tuple(toUInt64(1),toString('world'))) on local_node Executing query SELECT _part, * FROM mydb.tbl ORDER BY x on node2 Executing query system reload dictionary RedisSimple_flat_Int16_ on node [gw3] PASSED test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_complex[complex_key_direct] test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_complex[complex_key_hashed] Executing query system reload dictionaries on node Executing query select dictGetOrDefault('LocalClickHouse_complex_key_hashed_', 'Int32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt32(-77)) on local_node Executing query SELECT * FROM mydb.tbl ORDER BY x on node2 Executing query select dictGet('RedisSimple_flat_Int16_', 'Int16_', toUInt64(1)) on node Executing query select dictGet('MySQL_complex_key_hashed_', 'UInt8_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGetInt32OrDefault('LocalClickHouse_complex_key_hashed_', 'Int32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt32(-77)) on local_node Executing query DROP TABLE IF EXISTS tbl ON CLUSTER 'cluster3' SYNC on node1 [gw2] PASSED test_backup_restore_on_cluster/test.py::test_replicated_database_compare_parts Executing query select dictGetInt16('RedisSimple_flat_Int16_', 'Int16_', toUInt64(1)) on node Executing query select count() from system.tables where database = 'system' and name in ['query_log', 'text_log'] on node2 Executing query select dictGetUInt8('MySQL_complex_key_hashed_', 'UInt8_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGet('LocalClickHouse_complex_key_hashed_', 'Int64_', tuple(toUInt64(1),toString('world'))) on local_node Executing query select dictHas('RedisSimple_flat_Int16_', toUInt64(1)) on node Executing query DROP TABLE IF EXISTS tbl2 ON CLUSTER 'cluster3' SYNC on node1 Executing query rename table system.text_log to system.text_log_37_test on node2 Executing query select dictHas('MySQL_complex_key_hashed_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGetInt64('LocalClickHouse_complex_key_hashed_', 'Int64_', tuple(toUInt64(1),toString('world'))) on local_node Executing query select dictGetOrDefault('RedisSimple_flat_Int16_', 'Int16_', toUInt64(9999999), toInt16(-66)) on node Executing query rename table system.query_log to system.query_log_37_test on node2 Executing query DROP DATABASE IF EXISTS mydb ON CLUSTER 'cluster3' SYNC on node1 Executing query select dictGetOrDefault('MySQL_complex_key_hashed_', 'UInt8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt8(55)) on node Executing query select dictHas('LocalClickHouse_complex_key_hashed_', tuple(toUInt64(1),toString('world'))) on local_node Executing query system flush logs on node2 Executing query select dictGetInt16OrDefault('RedisSimple_flat_Int16_', 'Int16_', toUInt64(9999999), toInt16(-66)) on node Executing query select dictGetOrDefault('LocalClickHouse_complex_key_hashed_', 'Int64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt64(-88)) on local_node Executing query select dictGetUInt8OrDefault('MySQL_complex_key_hashed_', 'UInt8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt8(55)) on node Executing query select dictGet('RedisSimple_flat_Int16_', 'Int16_', toUInt64(2)) on node Executing query DROP DATABASE IF EXISTS mydb2 ON CLUSTER 'cluster3' SYNC on node1 Connection dropped: socket connection error: None Executing query select dictGetInt64OrDefault('LocalClickHouse_complex_key_hashed_', 'Int64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt64(-88)) on local_node Executing query select dictGet('MySQL_complex_key_hashed_', 'UInt16_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGetInt16('RedisSimple_flat_Int16_', 'Int16_', toUInt64(2)) on node Executing query DROP USER IF EXISTS u1, u2 ON CLUSTER 'cluster3' on node1 Executing query select dictGet('LocalClickHouse_complex_key_hashed_', 'UUID_', tuple(toUInt64(1),toString('world'))) on local_node Executing query select dictGetUInt16('MySQL_complex_key_hashed_', 'UInt16_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictHas('RedisSimple_flat_Int16_', toUInt64(2)) on node Executing query select count() from system.tables where database = 'system' and name in ['query_log', 'text_log'] on node2 test_backup_restore_on_cluster/test.py::test_replicated_database_with_special_macro_in_zk_path[database] Executing query CREATE DATABASE mydb ON CLUSTER 'cluster' ENGINE=Replicated('/clickhouse/databases/{database}','{shard}','{replica}') on node1 Executing query select dictGetUUID('LocalClickHouse_complex_key_hashed_', 'UUID_', tuple(toUInt64(1),toString('world'))) on local_node Executing query select dictHas('MySQL_complex_key_hashed_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGetOrDefault('RedisSimple_flat_Int16_', 'Int16_', toUInt64(9999999), toInt16(-66)) on node Executing query rename table system.text_log to system.text_log_38_test on node2 Executing query select dictGetInt16OrDefault('RedisSimple_flat_Int16_', 'Int16_', toUInt64(9999999), toInt16(-66)) on node Executing query CREATE TABLE mydb.tbl(x Int64) ENGINE=ReplicatedMergeTree ORDER BY x on node1 Executing query select dictHas('LocalClickHouse_complex_key_hashed_', tuple(toUInt64(1),toString('world'))) on local_node Executing query select dictGetOrDefault('MySQL_complex_key_hashed_', 'UInt16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt16(66)) on node Executing query rename table system.query_log to system.query_log_38_test on node2 Executing query system reload dictionary RedisSimple_hashed_Int16_ on node Executing query system flush logs on node2 Executing query select dictGetOrDefault('LocalClickHouse_complex_key_hashed_', 'UUID_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUUID('550e8400-0000-0000-0000-000000000000')) on local_node Executing query select dictGetUInt16OrDefault('MySQL_complex_key_hashed_', 'UInt16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt16(66)) on node Executing query INSERT INTO mydb.tbl VALUES (-3) on node1 Executing query select dictGet('RedisSimple_hashed_Int16_', 'Int16_', toUInt64(1)) on node Executing query select dictGetUUIDOrDefault('LocalClickHouse_complex_key_hashed_', 'UUID_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUUID('550e8400-0000-0000-0000-000000000000')) on local_node Executing query select dictGet('MySQL_complex_key_hashed_', 'UInt32_', tuple(toUInt64(1),toString('world'))) on node Executing query INSERT INTO mydb.tbl VALUES (1) on node1 Executing query select dictGetInt16('RedisSimple_hashed_Int16_', 'Int16_', toUInt64(1)) on node Executing query select dictGet('LocalClickHouse_complex_key_hashed_', 'Date_', tuple(toUInt64(1),toString('world'))) on local_node Executing query select dictGetUInt32('MySQL_complex_key_hashed_', 'UInt32_', tuple(toUInt64(1),toString('world'))) on node Executing query INSERT INTO mydb.tbl VALUES (10) on node1 Executing query select dictHas('RedisSimple_hashed_Int16_', toUInt64(1)) on node Executing query select dictGetDate('LocalClickHouse_complex_key_hashed_', 'Date_', tuple(toUInt64(1),toString('world'))) on local_node Executing query select dictHas('MySQL_complex_key_hashed_', tuple(toUInt64(1),toString('world'))) on node Executing query BACKUP DATABASE mydb ON CLUSTER 'cluster' TO Disk('backups', '15') on node1 Executing query select count() from system.tables where database = 'system' and name in ['query_log', 'text_log'] on node2 Executing query select dictGetOrDefault('RedisSimple_hashed_Int16_', 'Int16_', toUInt64(9999999), toInt16(-66)) on node Executing query select dictHas('LocalClickHouse_complex_key_hashed_', tuple(toUInt64(1),toString('world'))) on local_node Executing query select dictGetOrDefault('MySQL_complex_key_hashed_', 'UInt32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt32(77)) on node Executing query rename table system.text_log to system.text_log_39_test on node2 Executing query select dictGetInt16OrDefault('RedisSimple_hashed_Int16_', 'Int16_', toUInt64(9999999), toInt16(-66)) on node Executing query select dictGetOrDefault('LocalClickHouse_complex_key_hashed_', 'Date_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDate('2018-12-30')) on local_node Executing query select dictGetUInt32OrDefault('MySQL_complex_key_hashed_', 'UInt32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt32(77)) on node Executing query rename table system.query_log to system.query_log_39_test on node2 Executing query RESTORE DATABASE mydb AS mydb2 ON CLUSTER 'cluster' FROM Disk('backups', '15') on node1 Executing query select dictGet('RedisSimple_hashed_Int16_', 'Int16_', toUInt64(2)) on node Executing query select dictGetDateOrDefault('LocalClickHouse_complex_key_hashed_', 'Date_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDate('2018-12-30')) on local_node Executing query select dictGet('MySQL_complex_key_hashed_', 'UInt64_', tuple(toUInt64(1),toString('world'))) on node Executing query system flush logs on node2 Executing query select dictGetInt16('RedisSimple_hashed_Int16_', 'Int16_', toUInt64(2)) on node Executing query select dictGet('LocalClickHouse_complex_key_hashed_', 'DateTime_', tuple(toUInt64(1),toString('world'))) on local_node Executing query select dictGetUInt64('MySQL_complex_key_hashed_', 'UInt64_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGetDateTime('LocalClickHouse_complex_key_hashed_', 'DateTime_', tuple(toUInt64(1),toString('world'))) on local_node Executing query select dictHas('RedisSimple_hashed_Int16_', toUInt64(2)) on node Executing query select dictHas('MySQL_complex_key_hashed_', tuple(toUInt64(1),toString('world'))) on node Stderr: Container roottestdatabasebackup-gw1-resolver-1 Stopping Stderr: Container roottestdatabasebackup-gw1-instance-1 Stopping Stderr: Container roottestdatabasebackup-gw1-instance-1 Stopped Stderr: Container roottestdatabasebackup-gw1-minio1-1 Stopping Stderr: Container roottestdatabasebackup-gw1-minio1-1 Stopped Stderr: Container roottestdatabasebackup-gw1-resolver-1 Stopped Stderr: Container roottestdatabasebackup-gw1-proxy1-1 Stopping Stderr: Container roottestdatabasebackup-gw1-proxy2-1 Stopping Stderr: Container roottestdatabasebackup-gw1-proxy1-1 Stopped Stderr: Container roottestdatabasebackup-gw1-proxy2-1 Stopped Command:[bash -c [ -f /ClickHouse/tests/integration/test_database_backup/_instances-0-gw1/instance/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_database_backup/_instances-0-gw1/instance/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[docker compose --env-file /ClickHouse/tests/integration/test_database_backup/_instances-0-gw1/.env --project-name roottestdatabasebackup-gw1 --file /ClickHouse/tests/integration/test_database_backup/_instances-0-gw1/instance/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_minio.yml down --volumes] Executing query select dictHas('LocalClickHouse_complex_key_hashed_', tuple(toUInt64(1),toString('world'))) on local_node Executing query select dictGetOrDefault('MySQL_complex_key_hashed_', 'UInt64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt64(88)) on node Executing query INSERT INTO mydb.tbl VALUES (2) on node1 Executing query select dictGetOrDefault('RedisSimple_hashed_Int16_', 'Int16_', toUInt64(9999999), toInt16(-66)) on node Executing query select dictGetUInt64OrDefault('MySQL_complex_key_hashed_', 'UInt64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt64(88)) on node Executing query SYSTEM SYNC DATABASE REPLICA ON CLUSTER 'cluster' mydb2 on node1 Executing query select dictGetOrDefault('LocalClickHouse_complex_key_hashed_', 'DateTime_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDateTime('2018-12-30 00:00:00')) on local_node Executing query select dictGetInt16OrDefault('RedisSimple_hashed_Int16_', 'Int16_', toUInt64(9999999), toInt16(-66)) on node Executing query select count() from system.tables where database = 'system' and name in ['query_log', 'text_log'] on node2 Stderr: Container roottestdatabasebackup-gw1-instance-1 Stopping Stderr: Container roottestdatabasebackup-gw1-resolver-1 Stopping Stderr: Container roottestdatabasebackup-gw1-instance-1 Stopped Stderr: Container roottestdatabasebackup-gw1-instance-1 Removing Stderr: Container roottestdatabasebackup-gw1-resolver-1 Stopped Stderr: Container roottestdatabasebackup-gw1-resolver-1 Removing Stderr: Container roottestdatabasebackup-gw1-instance-1 Removed Stderr: Container roottestdatabasebackup-gw1-minio1-1 Stopping Stderr: Container roottestdatabasebackup-gw1-minio1-1 Stopped Stderr: Container roottestdatabasebackup-gw1-minio1-1 Removing Stderr: Container roottestdatabasebackup-gw1-resolver-1 Removed Stderr: Container roottestdatabasebackup-gw1-minio1-1 Removed Stderr: Container roottestdatabasebackup-gw1-proxy2-1 Stopping Stderr: Container roottestdatabasebackup-gw1-proxy1-1 Stopping Stderr: Container roottestdatabasebackup-gw1-proxy2-1 Stopped Stderr: Container roottestdatabasebackup-gw1-proxy2-1 Removing Stderr: Container roottestdatabasebackup-gw1-proxy1-1 Stopped Stderr: Container roottestdatabasebackup-gw1-proxy1-1 Removing Stderr: Container roottestdatabasebackup-gw1-proxy2-1 Removed Stderr: Container roottestdatabasebackup-gw1-proxy1-1 Removed Stderr: Volume roottestdatabasebackup-gw1_data1-1 Removing Stderr: Network roottestdatabasebackup-gw1_default Removing Stderr: Volume roottestdatabasebackup-gw1_data1-1 Removed Stderr: Network roottestdatabasebackup-gw1_default Removed Cleanup called Executing query select dictGetDateTimeOrDefault('LocalClickHouse_complex_key_hashed_', 'DateTime_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDateTime('2018-12-30 00:00:00')) on local_node Executing query system reload dictionary RedisSimple_cache_Int16_ on node Docker networks for project roottestdatabasebackup-gw1 are NETWORK ID NAME DRIVER SCOPE Executing query select dictGet('MySQL_complex_key_hashed_', 'Int8_', tuple(toUInt64(1),toString('world'))) on node Docker containers for project roottestdatabasebackup-gw1 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestdatabasebackup-gw1 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestdatabasebackup-gw1-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} No running containers for project: roottestdatabasebackup-gw1 Trying to prune unused networks... Trying to prune unused images... Command:[docker image prune -f] Executing query SYSTEM SYNC REPLICA ON CLUSTER 'cluster' mydb2.tbl on node1 Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Stdout:3 Command:[docker volume prune -f] Executing query rename table system.text_log to system.text_log_40_test on node2 Stdout:Total reclaimed space: 0B Volumes pruned: 3 Executing query select dictGet('RedisSimple_cache_Int16_', 'Int16_', toUInt64(1)) on node Executing query select dictGet('LocalClickHouse_complex_key_hashed_', 'String_', tuple(toUInt64(1),toString('world'))) on local_node Executing query select dictGetInt8('MySQL_complex_key_hashed_', 'Int8_', tuple(toUInt64(1),toString('world'))) on node Executing query rename table system.query_log to system.query_log_40_test on node2 Executing query SELECT * FROM mydb.tbl ORDER BY x on node1 Executing query select dictGetInt16('RedisSimple_cache_Int16_', 'Int16_', toUInt64(1)) on node Executing query select dictGetString('LocalClickHouse_complex_key_hashed_', 'String_', tuple(toUInt64(1),toString('world'))) on local_node run container_id:roottestasyncloaddatabases-gw7-node2-1 detach:False nothrow:True cmd: ['bash', '-c', 'ps -C clickhouse'] Command:[docker exec -u root roottestasyncloaddatabases-gw7-node2-1 bash -c ps -C clickhouse] Executing query select dictHas('MySQL_complex_key_hashed_', tuple(toUInt64(1),toString('world'))) on node Stdout: PID TTY TIME CMD Stdout: 2441 ? 00:00:37 clickhouse run container_id:roottestasyncloaddatabases-gw7-node2-1 detach:False nothrow:False cmd: ['bash', '-c', 'pkill clickhouse'] Command:[docker exec -u root roottestasyncloaddatabases-gw7-node2-1 bash -c pkill clickhouse] run container_id:roottestasyncloaddatabases-gw7-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw7-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:2441 Executing query SELECT * FROM mydb2.tbl ORDER BY x on node1 Executing query select dictHas('RedisSimple_cache_Int16_', toUInt64(1)) on node Executing query select dictHas('LocalClickHouse_complex_key_hashed_', tuple(toUInt64(1),toString('world'))) on local_node Executing query select dictGetOrDefault('MySQL_complex_key_hashed_', 'Int8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt8(-55)) on node Executing query SELECT * FROM mydb2.tbl ORDER BY x on node2 Executing query select dictGetOrDefault('RedisSimple_cache_Int16_', 'Int16_', toUInt64(9999999), toInt16(-66)) on node Executing query select dictGetOrDefault('LocalClickHouse_complex_key_hashed_', 'String_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toString('hi')) on local_node Executing query select dictGetInt8OrDefault('MySQL_complex_key_hashed_', 'Int8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt8(-55)) on node Executing query DROP TABLE IF EXISTS tbl ON CLUSTER 'cluster3' SYNC on node1 [gw2] PASSED test_backup_restore_on_cluster/test.py::test_replicated_database_with_special_macro_in_zk_path[database] Executing query select dictGetInt16OrDefault('RedisSimple_cache_Int16_', 'Int16_', toUInt64(9999999), toInt16(-66)) on node Executing query select dictGetStringOrDefault('LocalClickHouse_complex_key_hashed_', 'String_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toString('hi')) on local_node Executing query select dictGet('MySQL_complex_key_hashed_', 'Int16_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGet('LocalClickHouse_complex_key_hashed_', 'Float32_', tuple(toUInt64(1),toString('world'))) on local_node Executing query select dictGet('RedisSimple_cache_Int16_', 'Int16_', toUInt64(2)) on node Executing query DROP TABLE IF EXISTS tbl2 ON CLUSTER 'cluster3' SYNC on node1 run container_id:roottestasyncloaddatabases-gw7-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw7-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:2441 Executing query select dictGetInt16('MySQL_complex_key_hashed_', 'Int16_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGetFloat32('LocalClickHouse_complex_key_hashed_', 'Float32_', tuple(toUInt64(1),toString('world'))) on local_node Executing query select dictGetInt16('RedisSimple_cache_Int16_', 'Int16_', toUInt64(2)) on node Executing query select dictHas('MySQL_complex_key_hashed_', tuple(toUInt64(1),toString('world'))) on node Executing query DROP DATABASE IF EXISTS mydb ON CLUSTER 'cluster3' SYNC on node1 Executing query select dictHas('LocalClickHouse_complex_key_hashed_', tuple(toUInt64(1),toString('world'))) on local_node Executing query select dictHas('RedisSimple_cache_Int16_', toUInt64(2)) on node Executing query select dictGetOrDefault('MySQL_complex_key_hashed_', 'Int16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt16(-66)) on node Executing query DROP DATABASE IF EXISTS mydb2 ON CLUSTER 'cluster3' SYNC on node1 Executing query select dictGetOrDefault('LocalClickHouse_complex_key_hashed_', 'Float32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat32(555.11)) on local_node Executing query select dictGetOrDefault('RedisSimple_cache_Int16_', 'Int16_', toUInt64(9999999), toInt16(-66)) on node Stderr: backward Pulling Stderr: current Pulling Stderr: 4477f8fe99eb Pulling fs layer Stderr: 84f6d8efd27c Pulling fs layer Stderr: 4f4fb700ef54 Pulling fs layer Stderr: e5c63395e7a3 Pulling fs layer Stderr: 22e20d7710b3 Pulling fs layer Stderr: 184beec54d26 Pulling fs layer Stderr: 5add26bac827 Pulling fs layer Stderr: 6d305684be88 Pulling fs layer Stderr: cd532f345afc Pulling fs layer Stderr: 22e20d7710b3 Waiting Stderr: 184beec54d26 Waiting Stderr: e5c63395e7a3 Waiting Stderr: 5add26bac827 Waiting Stderr: 6d305684be88 Waiting Stderr: cd532f345afc Waiting Stderr: current Pulled Stderr: 84f6d8efd27c Downloading [> ] 163.8kB/15.37MB Stderr: 4f4fb700ef54 Downloading [==================================================>] 32B/32B Stderr: 4f4fb700ef54 Verifying Checksum Stderr: 4f4fb700ef54 Download complete Stderr: 4477f8fe99eb Downloading [> ] 275.9kB/27.51MB Stderr: e5c63395e7a3 Downloading [> ] 539.4kB/276.3MB Stderr: 84f6d8efd27c Verifying Checksum Stderr: 84f6d8efd27c Download complete Stderr: 22e20d7710b3 Downloading [=======> ] 3.646kB/24.12kB Stderr: 22e20d7710b3 Downloading [==================================================>] 24.12kB/24.12kB Stderr: 22e20d7710b3 Verifying Checksum Stderr: 22e20d7710b3 Download complete Stderr: 184beec54d26 Downloading [> ] 13.78kB/863.5kB Stderr: 184beec54d26 Verifying Checksum Stderr: 184beec54d26 Download complete Stderr: 5add26bac827 Downloading [==================================================>] 116B/116B Stderr: 5add26bac827 Download complete Stderr: 6d305684be88 Downloading [==================================================>] 362B/362B Stderr: 6d305684be88 Download complete Stderr: 4477f8fe99eb Download complete Stderr: cd532f345afc Downloading [==================================================>] 3.09kB/3.09kB Stderr: cd532f345afc Verifying Checksum Stderr: cd532f345afc Download complete Stderr: e5c63395e7a3 Downloading [=======> ] 41.1MB/276.3MB Stderr: 4477f8fe99eb Extracting [> ] 294.9kB/27.51MB Stderr: e5c63395e7a3 Downloading [===============> ] 83.99MB/276.3MB Stderr: 4477f8fe99eb Extracting [========> ] 4.719MB/27.51MB Stderr: e5c63395e7a3 Downloading [=======================> ] 129.7MB/276.3MB Stderr: 4477f8fe99eb Extracting [================> ] 8.847MB/27.51MB Stderr: e5c63395e7a3 Downloading [===============================> ] 176.5MB/276.3MB Stderr: 4477f8fe99eb Extracting [=======================> ] 12.98MB/27.51MB Stderr: e5c63395e7a3 Downloading [=======================================> ] 216.6MB/276.3MB Stderr: 4477f8fe99eb Extracting [===================================> ] 19.46MB/27.51MB Stderr: e5c63395e7a3 Downloading [==============================================> ] 258.9MB/276.3MB Stderr: e5c63395e7a3 Verifying Checksum Stderr: e5c63395e7a3 Download complete Stderr: 4477f8fe99eb Extracting [==========================================> ] 23.3MB/27.51MB Stderr: 4477f8fe99eb Extracting [===============================================> ] 26.25MB/27.51MB Stderr: 4477f8fe99eb Extracting [=================================================> ] 27.13MB/27.51MB Stderr: 4477f8fe99eb Extracting [==================================================>] 27.51MB/27.51MB Stderr: 4477f8fe99eb Pull complete Stderr: 84f6d8efd27c Extracting [> ] 163.8kB/15.37MB Stderr: 84f6d8efd27c Extracting [=========> ] 2.785MB/15.37MB Stderr: 84f6d8efd27c Extracting [============================> ] 8.847MB/15.37MB Stderr: 84f6d8efd27c Extracting [=====================================> ] 11.63MB/15.37MB Stderr: 84f6d8efd27c Extracting [=============================================> ] 14.09MB/15.37MB Stderr: 84f6d8efd27c Extracting [===============================================> ] 14.75MB/15.37MB Stderr: 84f6d8efd27c Extracting [================================================> ] 14.91MB/15.37MB Stderr: 84f6d8efd27c Extracting [==================================================>] 15.37MB/15.37MB Stderr: 84f6d8efd27c Pull complete Stderr: 4f4fb700ef54 Extracting [==================================================>] 32B/32B Stderr: 4f4fb700ef54 Extracting [==================================================>] 32B/32B Stderr: 4f4fb700ef54 Pull complete Stderr: e5c63395e7a3 Extracting [> ] 557.1kB/276.3MB Stderr: e5c63395e7a3 Extracting [=> ] 8.913MB/276.3MB Stderr: e5c63395e7a3 Extracting [===> ] 18.94MB/276.3MB Stderr: e5c63395e7a3 Extracting [====> ] 25.62MB/276.3MB Stderr: e5c63395e7a3 Extracting [=====> ] 32.87MB/276.3MB Stderr: e5c63395e7a3 Extracting [======> ] 37.88MB/276.3MB Stderr: e5c63395e7a3 Extracting [=======> ] 41.78MB/276.3MB Stderr: e5c63395e7a3 Extracting [========> ] 45.68MB/276.3MB Stderr: e5c63395e7a3 Extracting [=========> ] 51.25MB/276.3MB Stderr: e5c63395e7a3 Extracting [==========> ] 55.71MB/276.3MB Stderr: e5c63395e7a3 Extracting [===========> ] 61.28MB/276.3MB Stderr: e5c63395e7a3 Extracting [===========> ] 66.29MB/276.3MB Stderr: e5c63395e7a3 Extracting [============> ] 71.3MB/276.3MB Stderr: e5c63395e7a3 Extracting [=============> ] 76.32MB/276.3MB Stderr: e5c63395e7a3 Extracting [===============> ] 83.56MB/276.3MB Stderr: e5c63395e7a3 Extracting [================> ] 91.91MB/276.3MB Stderr: e5c63395e7a3 Extracting [==================> ] 101.9MB/276.3MB Stderr: e5c63395e7a3 Extracting [===================> ] 110.3MB/276.3MB Stderr: e5c63395e7a3 Extracting [=====================> ] 118.1MB/276.3MB Stderr: e5c63395e7a3 Extracting [======================> ] 125.9MB/276.3MB Stderr: e5c63395e7a3 Extracting [========================> ] 133.1MB/276.3MB Stderr: e5c63395e7a3 Extracting [========================> ] 135.4MB/276.3MB Stderr: e5c63395e7a3 Extracting [========================> ] 137MB/276.3MB Stderr: e5c63395e7a3 Extracting [=========================> ] 138.7MB/276.3MB Stderr: e5c63395e7a3 Extracting [==========================> ] 148.7MB/276.3MB Stderr: e5c63395e7a3 Extracting [=============================> ] 161MB/276.3MB Stderr: e5c63395e7a3 Extracting [==============================> ] 169.3MB/276.3MB Stderr: e5c63395e7a3 Extracting [===============================> ] 176MB/276.3MB Stderr: e5c63395e7a3 Extracting [=================================> ] 183.8MB/276.3MB Stderr: e5c63395e7a3 Extracting [==================================> ] 192.7MB/276.3MB Stderr: e5c63395e7a3 Extracting [====================================> ] 202.8MB/276.3MB Stderr: e5c63395e7a3 Extracting [======================================> ] 211.7MB/276.3MB Stderr: e5c63395e7a3 Extracting [=======================================> ] 218.4MB/276.3MB Stderr: e5c63395e7a3 Extracting [=========================================> ] 227.8MB/276.3MB Stderr: e5c63395e7a3 Extracting [==========================================> ] 235.1MB/276.3MB Stderr: e5c63395e7a3 Extracting [===========================================> ] 242.3MB/276.3MB Stderr: e5c63395e7a3 Extracting [=============================================> ] 251.8MB/276.3MB Stderr: e5c63395e7a3 Extracting [===============================================> ] 261.8MB/276.3MB Stderr: e5c63395e7a3 Extracting [================================================> ] 270.7MB/276.3MB Stderr: e5c63395e7a3 Extracting [=================================================> ] 275.7MB/276.3MB Stderr: e5c63395e7a3 Extracting [==================================================>] 276.3MB/276.3MB Stderr: e5c63395e7a3 Pull complete Stderr: 22e20d7710b3 Extracting [==================================================>] 24.12kB/24.12kB Stderr: 22e20d7710b3 Extracting [==================================================>] 24.12kB/24.12kB Stderr: 22e20d7710b3 Pull complete Stderr: 184beec54d26 Extracting [=> ] 32.77kB/863.5kB Stderr: 184beec54d26 Extracting [==================================================>] 863.5kB/863.5kB Stderr: 184beec54d26 Extracting [==================================================>] 863.5kB/863.5kB Stderr: 184beec54d26 Pull complete Stderr: 5add26bac827 Extracting [==================================================>] 116B/116B Stderr: 5add26bac827 Extracting [==================================================>] 116B/116B Stderr: 5add26bac827 Pull complete Stderr: 6d305684be88 Extracting [==================================================>] 362B/362B Stderr: 6d305684be88 Extracting [==================================================>] 362B/362B Stderr: 6d305684be88 Pull complete Stderr: cd532f345afc Extracting [==================================================>] 3.09kB/3.09kB Stderr: cd532f345afc Extracting [==================================================>] 3.09kB/3.09kB Stderr: cd532f345afc Pull complete Stderr: backward Pulled ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_analyzer_compatibility/_instances-0-gw4/.env --project-name roottestanalyzercompatibility-gw4 --file /ClickHouse/tests/integration/test_analyzer_compatibility/_instances-0-gw4/current/docker-compose.yml --file /ClickHouse/tests/integration/test_analyzer_compatibility/_instances-0-gw4/backward/docker-compose.yml up -d --no-recreate') Command:[docker compose --env-file /ClickHouse/tests/integration/test_analyzer_compatibility/_instances-0-gw4/.env --project-name roottestanalyzercompatibility-gw4 --file /ClickHouse/tests/integration/test_analyzer_compatibility/_instances-0-gw4/current/docker-compose.yml --file /ClickHouse/tests/integration/test_analyzer_compatibility/_instances-0-gw4/backward/docker-compose.yml up -d --no-recreate] Executing query select dictGetInt16OrDefault('MySQL_complex_key_hashed_', 'Int16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt16(-66)) on node Executing query select dictGetInt16OrDefault('RedisSimple_cache_Int16_', 'Int16_', toUInt64(9999999), toInt16(-66)) on node run container_id:roottestasyncloaddatabases-gw7-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw7-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select dictGetFloat32OrDefault('LocalClickHouse_complex_key_hashed_', 'Float32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat32(555.11)) on local_node Stdout:2441 Executing query DROP USER IF EXISTS u1, u2 ON CLUSTER 'cluster3' on node1 Executing query select dictGet('MySQL_complex_key_hashed_', 'Int32_', tuple(toUInt64(1),toString('world'))) on node Executing query system reload dictionary RedisSimple_direct_Int16_ on node Executing query select dictGet('LocalClickHouse_complex_key_hashed_', 'Float64_', tuple(toUInt64(1),toString('world'))) on local_node Connection dropped: socket connection error: None Executing query select dictGetInt32('MySQL_complex_key_hashed_', 'Int32_', tuple(toUInt64(1),toString('world'))) on node test_backup_restore_on_cluster/test.py::test_replicated_database_with_special_macro_in_zk_path[uuid] Executing query CREATE DATABASE mydb ON CLUSTER 'cluster' ENGINE=Replicated('/clickhouse/databases/{uuid}','{shard}','{replica}') on node1 Executing query select dictGet('RedisSimple_direct_Int16_', 'Int16_', toUInt64(1)) on node Executing query select dictGetFloat64('LocalClickHouse_complex_key_hashed_', 'Float64_', tuple(toUInt64(1),toString('world'))) on local_node Stderr: Network roottestanalyzercompatibility-gw4_default Creating Stderr: Network roottestanalyzercompatibility-gw4_default Created Stderr: Container roottestanalyzercompatibility-gw4-backward-1 Creating Stderr: Container roottestanalyzercompatibility-gw4-current-1 Creating Stderr: Container roottestanalyzercompatibility-gw4-backward-1 Created Stderr: Container roottestanalyzercompatibility-gw4-current-1 Created Stderr: Container roottestanalyzercompatibility-gw4-backward-1 Starting Stderr: Container roottestanalyzercompatibility-gw4-current-1 Starting Stderr: Container roottestanalyzercompatibility-gw4-current-1 Started Stderr: Container roottestanalyzercompatibility-gw4-backward-1 Started ClickHouse instance created get_instance_ip instance_name=current http://localhost:None "GET /v1.46/containers/roottestanalyzercompatibility-gw4-current-1/json HTTP/1.1" 200 None get_instance_ip instance_name=current http://localhost:None "GET /v1.46/containers/roottestanalyzercompatibility-gw4-current-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in current, ip: 172.16.1.3... http://localhost:None "GET /v1.46/containers/roottestanalyzercompatibility-gw4-current-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/192aa1c12ae470d0d890f4e870cdf398bb0a0fe4049774833b14d4ef1aabb7f3/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/192aa1c12ae470d0d890f4e870cdf398bb0a0fe4049774833b14d4ef1aabb7f3/json HTTP/1.1" 200 None Executing query select dictHas('MySQL_complex_key_hashed_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGetInt16('RedisSimple_direct_Int16_', 'Int16_', toUInt64(1)) on node Executing query select dictHas('LocalClickHouse_complex_key_hashed_', tuple(toUInt64(1),toString('world'))) on local_node http://localhost:None "GET /v1.46/containers/192aa1c12ae470d0d890f4e870cdf398bb0a0fe4049774833b14d4ef1aabb7f3/json HTTP/1.1" 200 None Executing query CREATE TABLE mydb.tbl(x Int64) ENGINE=ReplicatedMergeTree ORDER BY x on node1 http://localhost:None "GET /v1.46/containers/192aa1c12ae470d0d890f4e870cdf398bb0a0fe4049774833b14d4ef1aabb7f3/json HTTP/1.1" 200 None run container_id:roottestasyncloaddatabases-gw7-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw7-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:2441 http://localhost:None "GET /v1.46/containers/192aa1c12ae470d0d890f4e870cdf398bb0a0fe4049774833b14d4ef1aabb7f3/json HTTP/1.1" 200 None Executing query select dictHas('RedisSimple_direct_Int16_', toUInt64(1)) on node Executing query select dictGetOrDefault('LocalClickHouse_complex_key_hashed_', 'Float64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat64(777.11)) on local_node Executing query select dictGetOrDefault('MySQL_complex_key_hashed_', 'Int32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt32(-77)) on node http://localhost:None "GET /v1.46/containers/192aa1c12ae470d0d890f4e870cdf398bb0a0fe4049774833b14d4ef1aabb7f3/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/192aa1c12ae470d0d890f4e870cdf398bb0a0fe4049774833b14d4ef1aabb7f3/json HTTP/1.1" 200 None Executing query INSERT INTO mydb.tbl VALUES (-3) on node1 http://localhost:None "GET /v1.46/containers/192aa1c12ae470d0d890f4e870cdf398bb0a0fe4049774833b14d4ef1aabb7f3/json HTTP/1.1" 200 None Executing query select dictGetFloat64OrDefault('LocalClickHouse_complex_key_hashed_', 'Float64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat64(777.11)) on local_node Executing query select dictGetOrDefault('RedisSimple_direct_Int16_', 'Int16_', toUInt64(9999999), toInt16(-66)) on node Executing query select dictGetInt32OrDefault('MySQL_complex_key_hashed_', 'Int32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt32(-77)) on node http://localhost:None "GET /v1.46/containers/192aa1c12ae470d0d890f4e870cdf398bb0a0fe4049774833b14d4ef1aabb7f3/json HTTP/1.1" 200 None Executing query INSERT INTO mydb.tbl VALUES (1) on node1 http://localhost:None "GET /v1.46/containers/192aa1c12ae470d0d890f4e870cdf398bb0a0fe4049774833b14d4ef1aabb7f3/json HTTP/1.1" 200 None Executing query select dictGetInt16OrDefault('RedisSimple_direct_Int16_', 'Int16_', toUInt64(9999999), toInt16(-66)) on node http://localhost:None "GET /v1.46/containers/192aa1c12ae470d0d890f4e870cdf398bb0a0fe4049774833b14d4ef1aabb7f3/json HTTP/1.1" 200 None Executing query select dictGet('MySQL_complex_key_hashed_', 'Int64_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGet('LocalClickHouse_complex_key_hashed_', 'UInt8_', tuple(toUInt64(2),toString('qwerty2'))) on local_node http://localhost:None "GET /v1.46/containers/192aa1c12ae470d0d890f4e870cdf398bb0a0fe4049774833b14d4ef1aabb7f3/json HTTP/1.1" 200 None Executing query INSERT INTO mydb.tbl VALUES (10) on node1 http://localhost:None "GET /v1.46/containers/192aa1c12ae470d0d890f4e870cdf398bb0a0fe4049774833b14d4ef1aabb7f3/json HTTP/1.1" 200 None Executing query select dictGet('RedisSimple_direct_Int16_', 'Int16_', toUInt64(2)) on node Executing query select dictGetUInt8('LocalClickHouse_complex_key_hashed_', 'UInt8_', tuple(toUInt64(2),toString('qwerty2'))) on local_node http://localhost:None "GET /v1.46/containers/192aa1c12ae470d0d890f4e870cdf398bb0a0fe4049774833b14d4ef1aabb7f3/json HTTP/1.1" 200 None Executing query select dictGetInt64('MySQL_complex_key_hashed_', 'Int64_', tuple(toUInt64(1),toString('world'))) on node run container_id:roottestasyncloaddatabases-gw7-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw7-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:2441 http://localhost:None "GET /v1.46/containers/192aa1c12ae470d0d890f4e870cdf398bb0a0fe4049774833b14d4ef1aabb7f3/json HTTP/1.1" 200 None Executing query BACKUP DATABASE mydb ON CLUSTER 'cluster' TO Disk('backups', '16') on node1 Executing query select dictGetInt16('RedisSimple_direct_Int16_', 'Int16_', toUInt64(2)) on node http://localhost:None "GET /v1.46/containers/192aa1c12ae470d0d890f4e870cdf398bb0a0fe4049774833b14d4ef1aabb7f3/json HTTP/1.1" 200 None Executing query select dictHas('LocalClickHouse_complex_key_hashed_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select dictHas('MySQL_complex_key_hashed_', tuple(toUInt64(1),toString('world'))) on node http://localhost:None "GET /v1.46/containers/192aa1c12ae470d0d890f4e870cdf398bb0a0fe4049774833b14d4ef1aabb7f3/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/192aa1c12ae470d0d890f4e870cdf398bb0a0fe4049774833b14d4ef1aabb7f3/json HTTP/1.1" 200 None Executing query select dictHas('RedisSimple_direct_Int16_', toUInt64(2)) on node Executing query select dictGetOrDefault('LocalClickHouse_complex_key_hashed_', 'UInt8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt8(55)) on local_node http://localhost:None "GET /v1.46/containers/192aa1c12ae470d0d890f4e870cdf398bb0a0fe4049774833b14d4ef1aabb7f3/json HTTP/1.1" 200 None Executing query select dictGetOrDefault('MySQL_complex_key_hashed_', 'Int64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt64(-88)) on node http://localhost:None "GET /v1.46/containers/192aa1c12ae470d0d890f4e870cdf398bb0a0fe4049774833b14d4ef1aabb7f3/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/192aa1c12ae470d0d890f4e870cdf398bb0a0fe4049774833b14d4ef1aabb7f3/json HTTP/1.1" 200 None Executing query select dictGetUInt8OrDefault('LocalClickHouse_complex_key_hashed_', 'UInt8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt8(55)) on local_node Executing query select dictGetOrDefault('RedisSimple_direct_Int16_', 'Int16_', toUInt64(9999999), toInt16(-66)) on node Executing query select dictGetInt64OrDefault('MySQL_complex_key_hashed_', 'Int64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt64(-88)) on node http://localhost:None "GET /v1.46/containers/192aa1c12ae470d0d890f4e870cdf398bb0a0fe4049774833b14d4ef1aabb7f3/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/192aa1c12ae470d0d890f4e870cdf398bb0a0fe4049774833b14d4ef1aabb7f3/json HTTP/1.1" 200 None Executing query RESTORE DATABASE mydb AS mydb2 ON CLUSTER 'cluster' FROM Disk('backups', '16') on node1 ClickHouse current started get_instance_ip instance_name=backward http://localhost:None "GET /v1.46/containers/roottestanalyzercompatibility-gw4-backward-1/json HTTP/1.1" 200 None get_instance_ip instance_name=backward http://localhost:None "GET /v1.46/containers/roottestanalyzercompatibility-gw4-backward-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in backward, ip: 172.16.1.2... http://localhost:None "GET /v1.46/containers/roottestanalyzercompatibility-gw4-backward-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/1e1f61243701c4353ee64572db21e33b69d6fb6e23c783b205227cc7aac275d8/json HTTP/1.1" 200 None ClickHouse backward started Executing query SYSTEM FLUSH LOGS on current run container_id:roottestasyncloaddatabases-gw7-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw7-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select dictGet('LocalClickHouse_complex_key_hashed_', 'UInt16_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select dictGetInt16OrDefault('RedisSimple_direct_Int16_', 'Int16_', toUInt64(9999999), toInt16(-66)) on node run container_id:roottestasyncloaddatabases-gw7-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw7-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select dictGet('MySQL_complex_key_hashed_', 'UUID_', tuple(toUInt64(1),toString('world'))) on node No clickhouse process running. Start new one. http://localhost:None "POST /v1.46/containers/roottestasyncloaddatabases-gw7-node2-1/exec HTTP/1.1" 201 74 http://localhost:None "POST /v1.46/exec/eb2cf59e35cf9590bdc9de47b109ef1f578b479aefafdd7d359ed72811d88663/start HTTP/1.1" 200 0 http://localhost:None "GET /v1.46/exec/eb2cf59e35cf9590bdc9de47b109ef1f578b479aefafdd7d359ed72811d88663/json HTTP/1.1" 200 586 Executing query select dictGetUInt16('LocalClickHouse_complex_key_hashed_', 'UInt16_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query system reload dictionary RedisHash_complex_key_hashed_Int16_ on node Executing query select dictGetUUID('MySQL_complex_key_hashed_', 'UUID_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictHas('LocalClickHouse_complex_key_hashed_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select dictGet('RedisHash_complex_key_hashed_Int16_', 'Int16_', tuple(toUInt64(1),toString('world'))) on node Executing query INSERT INTO mydb.tbl VALUES (2) on node1 Executing query select dictHas('MySQL_complex_key_hashed_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGetOrDefault('LocalClickHouse_complex_key_hashed_', 'UInt16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt16(66)) on local_node Executing query select dictGetInt16('RedisHash_complex_key_hashed_Int16_', 'Int16_', tuple(toUInt64(1),toString('world'))) on node Executing query SYSTEM SYNC DATABASE REPLICA ON CLUSTER 'cluster' mydb2 on node1 Executing query select dictGetOrDefault('MySQL_complex_key_hashed_', 'UUID_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUUID('550e8400-0000-0000-0000-000000000000')) on node Executing query SYSTEM FLUSH LOGS on backward Executing query select dictGetUInt16OrDefault('LocalClickHouse_complex_key_hashed_', 'UInt16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt16(66)) on local_node run container_id:roottestasyncloaddatabases-gw7-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw7-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select dictHas('RedisHash_complex_key_hashed_Int16_', tuple(toUInt64(1),toString('world'))) on node Stdout:3271 Clickhouse process running. run container_id:roottestasyncloaddatabases-gw7-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw7-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select dictGetUUIDOrDefault('MySQL_complex_key_hashed_', 'UUID_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUUID('550e8400-0000-0000-0000-000000000000')) on node Stdout:3271 Executing query select 20 on node2 Executing query SYSTEM SYNC REPLICA ON CLUSTER 'cluster' mydb2.tbl on node1 Executing query select dictGet('LocalClickHouse_complex_key_hashed_', 'UInt32_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query SELECT name FROM clusterAllReplicas('test_cluster_mixed', system.tables); on current Executing query select dictGetOrDefault('RedisHash_complex_key_hashed_Int16_', 'Int16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt16(-66)) on node Executing query select dictGet('MySQL_complex_key_hashed_', 'Date_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGetUInt32('LocalClickHouse_complex_key_hashed_', 'UInt32_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select dictGetInt16OrDefault('RedisHash_complex_key_hashed_Int16_', 'Int16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt16(-66)) on node Executing query SYSTEM FLUSH LOGS on current Executing query select dictGetDate('MySQL_complex_key_hashed_', 'Date_', tuple(toUInt64(1),toString('world'))) on node Executing query SELECT * FROM mydb.tbl ORDER BY x on node1 Executing query select 20 on node2 Executing query select dictHas('LocalClickHouse_complex_key_hashed_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select dictGet('RedisHash_complex_key_hashed_Int16_', 'Int16_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictHas('MySQL_complex_key_hashed_', tuple(toUInt64(1),toString('world'))) on node Executing query SELECT * FROM mydb2.tbl ORDER BY x on node1 Executing query select dictGetOrDefault('LocalClickHouse_complex_key_hashed_', 'UInt32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt32(77)) on local_node Executing query select dictGetInt16('RedisHash_complex_key_hashed_Int16_', 'Int16_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGetOrDefault('MySQL_complex_key_hashed_', 'Date_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDate('2018-12-30')) on node Executing query SELECT * FROM mydb2.tbl ORDER BY x on node2 Executing query select 20 on node2 Executing query select dictGetUInt32OrDefault('LocalClickHouse_complex_key_hashed_', 'UInt32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt32(77)) on local_node Executing query select dictHas('RedisHash_complex_key_hashed_Int16_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGetDateOrDefault('MySQL_complex_key_hashed_', 'Date_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDate('2018-12-30')) on node Executing query SYSTEM FLUSH LOGS on backward Executing query DROP TABLE IF EXISTS tbl ON CLUSTER 'cluster3' SYNC on node1 [gw2] PASSED test_backup_restore_on_cluster/test.py::test_replicated_database_with_special_macro_in_zk_path[uuid] Executing query select dictGetOrDefault('RedisHash_complex_key_hashed_Int16_', 'Int16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt16(-66)) on node Executing query select dictGet('LocalClickHouse_complex_key_hashed_', 'UInt64_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select dictGet('MySQL_complex_key_hashed_', 'DateTime_', tuple(toUInt64(1),toString('world'))) on node Executing query SELECT hostname() AS h, getSetting('allow_experimental_analyzer') FROM clusterAllReplicas('test_cluster_mixed', system.one) ORDER BY h; on current Executing query select dictGetInt16OrDefault('RedisHash_complex_key_hashed_Int16_', 'Int16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt16(-66)) on node Executing query select dictGetUInt64('LocalClickHouse_complex_key_hashed_', 'UInt64_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query DROP TABLE IF EXISTS tbl2 ON CLUSTER 'cluster3' SYNC on node1 Executing query select 20 on node2 Executing query select dictGetDateTime('MySQL_complex_key_hashed_', 'DateTime_', tuple(toUInt64(1),toString('world'))) on node Executing query SELECT DISTINCT Settings['allow_experimental_analyzer'] FROM system.query_log WHERE initial_query_id = '9c4bc824-ccfd-4819-a5b7-d7cdf6612b81'; on backward Executing query system reload dictionary RedisHash_complex_key_cache_Int16_ on node Executing query select dictHas('LocalClickHouse_complex_key_hashed_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select dictHas('MySQL_complex_key_hashed_', tuple(toUInt64(1),toString('world'))) on node Executing query SELECT name FROM clusterAllReplicas('test_cluster_mixed', system.tables) on backward Executing query DROP DATABASE IF EXISTS mydb ON CLUSTER 'cluster3' SYNC on node1 Executing query select dictGet('RedisHash_complex_key_cache_Int16_', 'Int16_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGetOrDefault('LocalClickHouse_complex_key_hashed_', 'UInt64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt64(88)) on local_node Executing query select dictGetOrDefault('MySQL_complex_key_hashed_', 'DateTime_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDateTime('2018-12-30 00:00:00')) on node Executing query SYSTEM FLUSH LOGS on current Executing query select 20 on node2 Executing query select dictGetInt16('RedisHash_complex_key_cache_Int16_', 'Int16_', tuple(toUInt64(1),toString('world'))) on node Executing query DROP DATABASE IF EXISTS mydb2 ON CLUSTER 'cluster3' SYNC on node1 Executing query select dictGetUInt64OrDefault('LocalClickHouse_complex_key_hashed_', 'UInt64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt64(88)) on local_node Executing query select dictGetDateTimeOrDefault('MySQL_complex_key_hashed_', 'DateTime_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDateTime('2018-12-30 00:00:00')) on node Executing query select dictHas('RedisHash_complex_key_cache_Int16_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGet('LocalClickHouse_complex_key_hashed_', 'Int8_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select dictGet('MySQL_complex_key_hashed_', 'String_', tuple(toUInt64(1),toString('world'))) on node Executing query DROP USER IF EXISTS u1, u2 ON CLUSTER 'cluster3' on node1 Executing query select dictGetOrDefault('RedisHash_complex_key_cache_Int16_', 'Int16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt16(-66)) on node Executing query select dictGetInt8('LocalClickHouse_complex_key_hashed_', 'Int8_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select dictGetString('MySQL_complex_key_hashed_', 'String_', tuple(toUInt64(1),toString('world'))) on node Executing query SYSTEM FLUSH LOGS on backward Executing query CREATE TABLE tbl ON CLUSTER 'cluster' (x UInt8, y String) ENGINE=ReplicatedMergeTree('/clickhouse/tables/tbl/', '{replica}')ORDER BY x on node1 test_backup_restore_on_cluster/test.py::test_replicated_table Executing query select dictGetInt16OrDefault('RedisHash_complex_key_cache_Int16_', 'Int16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt16(-66)) on node Executing query select dictHas('LocalClickHouse_complex_key_hashed_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select dictHas('MySQL_complex_key_hashed_', tuple(toUInt64(1),toString('world'))) on node Executing query SELECT hostname() AS h, getSetting('allow_experimental_analyzer') FROM clusterAllReplicas('test_cluster_mixed', system.one) ORDER BY h; on backward Executing query select dictGet('RedisHash_complex_key_cache_Int16_', 'Int16_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGetOrDefault('LocalClickHouse_complex_key_hashed_', 'Int8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt8(-55)) on local_node Executing query select dictGetOrDefault('MySQL_complex_key_hashed_', 'String_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toString('hi')) on node Executing query INSERT INTO tbl VALUES (1, 'Don''t') on node1 Executing query SELECT DISTINCT Settings['allow_experimental_analyzer'] FROM clusterAllReplicas('test_cluster_mixed', system.query_log) WHERE initial_query_id = '09ba9e0c-2c22-4a3f-92d0-637b79edbe1a'; on backward Executing query select dictGetInt16('RedisHash_complex_key_cache_Int16_', 'Int16_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGetInt8OrDefault('LocalClickHouse_complex_key_hashed_', 'Int8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt8(-55)) on local_node Executing query select dictGetStringOrDefault('MySQL_complex_key_hashed_', 'String_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toString('hi')) on node Executing query INSERT INTO tbl VALUES (2, 'count') on node2 Executing query select dictHas('RedisHash_complex_key_cache_Int16_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query SELECT name FROM clusterAllReplicas('test_cluster_mixed', system.tables) SETTINGS enable_analyzer = 1; on current Executing query INSERT INTO tbl SETTINGS async_insert=true VALUES (3, 'your') on node1 Executing query select dictGet('LocalClickHouse_complex_key_hashed_', 'Int16_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select dictGet('MySQL_complex_key_hashed_', 'Float32_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGetOrDefault('RedisHash_complex_key_cache_Int16_', 'Int16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt16(-66)) on node Executing query drop table if exists system.text_log_1_test on node2 Executing query SYSTEM FLUSH LOGS on current Executing query select dictGetInt16('LocalClickHouse_complex_key_hashed_', 'Int16_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query INSERT INTO tbl SETTINGS async_insert=true VALUES (4, 'chickens') on node2 Executing query select dictGetFloat32('MySQL_complex_key_hashed_', 'Float32_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGetInt16OrDefault('RedisHash_complex_key_cache_Int16_', 'Int16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt16(-66)) on node Executing query drop table if exists system.query_log_1_test on node2 Executing query select dictHas('LocalClickHouse_complex_key_hashed_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select dictHas('MySQL_complex_key_hashed_', tuple(toUInt64(1),toString('world'))) on node Executing query SYSTEM SYNC REPLICA ON CLUSTER 'cluster' tbl on node1 Executing query drop table if exists system.text_log_2_test on node2 Executing query system reload dictionary RedisHash_complex_key_direct_Int16_ on node Executing query select dictGetOrDefault('LocalClickHouse_complex_key_hashed_', 'Int16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt16(-66)) on local_node Executing query select dictGetOrDefault('MySQL_complex_key_hashed_', 'Float32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat32(555.11)) on node Executing query drop table if exists system.query_log_2_test on node2 Executing query select dictGet('RedisHash_complex_key_direct_Int16_', 'Int16_', tuple(toUInt64(1),toString('world'))) on node Executing query BACKUP TABLE tbl ON CLUSTER 'cluster' TO Disk('backups', '17') SETTINGS replica_num=1 on node1 Executing query SYSTEM FLUSH LOGS on backward Executing query select dictGetInt16OrDefault('LocalClickHouse_complex_key_hashed_', 'Int16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt16(-66)) on local_node Executing query select dictGetFloat32OrDefault('MySQL_complex_key_hashed_', 'Float32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat32(555.11)) on node Executing query drop table if exists system.text_log_3_test on node2 Executing query select dictGetInt16('RedisHash_complex_key_direct_Int16_', 'Int16_', tuple(toUInt64(1),toString('world'))) on node Executing query SELECT DISTINCT Settings['allow_experimental_analyzer'] FROM system.query_log WHERE initial_query_id = 'f84534f9-3bd1-441e-9cf1-1957261c6422'; on current Executing query select dictGet('LocalClickHouse_complex_key_hashed_', 'Int32_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query drop table if exists system.query_log_3_test on node2 Executing query select dictGet('MySQL_complex_key_hashed_', 'Float64_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictHas('RedisHash_complex_key_direct_Int16_', tuple(toUInt64(1),toString('world'))) on node Executing query DROP TABLE tbl ON CLUSTER 'cluster' SYNC on node1 Executing query drop table if exists system.text_log_4_test on node2 Command:[docker compose --env-file /ClickHouse/tests/integration/test_analyzer_compatibility/_instances-0-gw4/.env --project-name roottestanalyzercompatibility-gw4 --file /ClickHouse/tests/integration/test_analyzer_compatibility/_instances-0-gw4/current/docker-compose.yml --file /ClickHouse/tests/integration/test_analyzer_compatibility/_instances-0-gw4/backward/docker-compose.yml stop --timeout 20] [gw4] PASSED test_analyzer_compatibility/test.py::test_two_new_versions Executing query select dictGetInt32('LocalClickHouse_complex_key_hashed_', 'Int32_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select dictGetOrDefault('RedisHash_complex_key_direct_Int16_', 'Int16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt16(-66)) on node Executing query select dictGetFloat64('MySQL_complex_key_hashed_', 'Float64_', tuple(toUInt64(1),toString('world'))) on node Executing query drop table if exists system.query_log_4_test on node2 Executing query RESTORE TABLE tbl ON CLUSTER 'cluster' FROM Disk('backups', '17') on node2 Executing query select dictHas('LocalClickHouse_complex_key_hashed_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select dictGetInt16OrDefault('RedisHash_complex_key_direct_Int16_', 'Int16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt16(-66)) on node Executing query select dictHas('MySQL_complex_key_hashed_', tuple(toUInt64(1),toString('world'))) on node Executing query drop table if exists system.text_log_5_test on node2 Executing query drop table if exists system.query_log_5_test on node2 Executing query select dictGetOrDefault('LocalClickHouse_complex_key_hashed_', 'Int32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt32(-77)) on local_node Executing query select dictGet('RedisHash_complex_key_direct_Int16_', 'Int16_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGetOrDefault('MySQL_complex_key_hashed_', 'Float64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat64(777.11)) on node Executing query drop table if exists system.text_log_6_test on node2 Executing query select dictGetFloat64OrDefault('MySQL_complex_key_hashed_', 'Float64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat64(777.11)) on node Executing query select dictGetInt32OrDefault('LocalClickHouse_complex_key_hashed_', 'Int32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt32(-77)) on local_node Executing query select dictGetInt16('RedisHash_complex_key_direct_Int16_', 'Int16_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query SYSTEM SYNC REPLICA ON CLUSTER 'cluster' tbl on node1 Executing query drop table if exists system.query_log_6_test on node2 Executing query select dictGet('MySQL_complex_key_hashed_', 'UInt8_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGet('LocalClickHouse_complex_key_hashed_', 'Int64_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select dictHas('RedisHash_complex_key_direct_Int16_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query drop table if exists system.text_log_7_test on node2 Executing query SELECT * FROM tbl ORDER BY x on node2 Executing query drop table if exists system.query_log_7_test on node2 Executing query select dictGetUInt8('MySQL_complex_key_hashed_', 'UInt8_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGetInt64('LocalClickHouse_complex_key_hashed_', 'Int64_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select dictGetOrDefault('RedisHash_complex_key_direct_Int16_', 'Int16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt16(-66)) on node Executing query drop table if exists system.text_log_8_test on node2 Executing query SELECT * FROM tbl ORDER BY x on node1 Executing query select dictHas('MySQL_complex_key_hashed_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictHas('LocalClickHouse_complex_key_hashed_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select dictGetInt16OrDefault('RedisHash_complex_key_direct_Int16_', 'Int16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt16(-66)) on node Executing query drop table if exists system.query_log_8_test on node2 Executing query select dictGetOrDefault('MySQL_complex_key_hashed_', 'UInt8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt8(55)) on node Executing query select dictGetOrDefault('LocalClickHouse_complex_key_hashed_', 'Int64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt64(-88)) on local_node Executing query DROP TABLE IF EXISTS tbl ON CLUSTER 'cluster3' SYNC on node1 [gw2] PASSED test_backup_restore_on_cluster/test.py::test_replicated_table Executing query drop table if exists system.text_log_9_test on node2 Executing query system reload dictionaries on node Executing query select dictGetUInt8OrDefault('MySQL_complex_key_hashed_', 'UInt8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt8(55)) on node Executing query drop table if exists system.query_log_9_test on node2 Executing query select dictGetInt64OrDefault('LocalClickHouse_complex_key_hashed_', 'Int64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt64(-88)) on local_node [gw0] PASSED test_dictionaries_redis/test.py::test_redis_dictionaries[Int16_] test_dictionaries_redis/test.py::test_redis_dictionaries[Int32_] Run test with id: 6 Executing query system reload dictionaries on node Executing query DROP TABLE IF EXISTS tbl2 ON CLUSTER 'cluster3' SYNC on node1 Executing query drop table if exists system.text_log_10_test on node2 Executing query select dictGet('MySQL_complex_key_hashed_', 'UInt16_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGet('LocalClickHouse_complex_key_hashed_', 'UUID_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query system reload dictionary RedisSimple_flat_Int32_ on node Executing query drop table if exists system.query_log_10_test on node2 Executing query DROP DATABASE IF EXISTS mydb ON CLUSTER 'cluster3' SYNC on node1 Executing query select dictGetUInt16('MySQL_complex_key_hashed_', 'UInt16_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGet('RedisSimple_flat_Int32_', 'Int32_', toUInt64(1)) on node Executing query select dictGetUUID('LocalClickHouse_complex_key_hashed_', 'UUID_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query drop table if exists system.text_log_11_test on node2 Executing query select dictGetInt32('RedisSimple_flat_Int32_', 'Int32_', toUInt64(1)) on node Executing query select dictHas('MySQL_complex_key_hashed_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query drop table if exists system.query_log_11_test on node2 Executing query DROP DATABASE IF EXISTS mydb2 ON CLUSTER 'cluster3' SYNC on node1 Executing query select dictHas('LocalClickHouse_complex_key_hashed_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select dictHas('RedisSimple_flat_Int32_', toUInt64(1)) on node Executing query drop table if exists system.text_log_12_test on node2 Executing query select dictGetOrDefault('MySQL_complex_key_hashed_', 'UInt16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt16(66)) on node Executing query select dictGetOrDefault('LocalClickHouse_complex_key_hashed_', 'UUID_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUUID('550e8400-0000-0000-0000-000000000000')) on local_node Executing query DROP USER IF EXISTS u1, u2 ON CLUSTER 'cluster3' on node1 Executing query drop table if exists system.query_log_12_test on node2 Executing query select dictGetOrDefault('RedisSimple_flat_Int32_', 'Int32_', toUInt64(9999999), toInt32(-77)) on node Executing query select dictGetUUIDOrDefault('LocalClickHouse_complex_key_hashed_', 'UUID_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUUID('550e8400-0000-0000-0000-000000000000')) on local_node Executing query select dictGetUInt16OrDefault('MySQL_complex_key_hashed_', 'UInt16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt16(66)) on node Executing query drop table if exists system.text_log_13_test on node2 test_backup_restore_on_cluster/test.py::test_replicated_table_after_alters Executing query CREATE TABLE tbl ON CLUSTER 'cluster' (x Int32) ENGINE=ReplicatedMergeTree('/clickhouse/tables/tbl/', '{replica}')ORDER BY x on node1 Executing query select dictGetInt32OrDefault('RedisSimple_flat_Int32_', 'Int32_', toUInt64(9999999), toInt32(-77)) on node Executing query select dictGet('LocalClickHouse_complex_key_hashed_', 'Date_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select dictGet('MySQL_complex_key_hashed_', 'UInt32_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query drop table if exists system.query_log_13_test on node2 Executing query select dictGet('RedisSimple_flat_Int32_', 'Int32_', toUInt64(2)) on node Executing query drop table if exists system.text_log_14_test on node2 Executing query select dictGetDate('LocalClickHouse_complex_key_hashed_', 'Date_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select dictGetUInt32('MySQL_complex_key_hashed_', 'UInt32_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query INSERT INTO tbl VALUES (1) on node1 Executing query drop table if exists system.query_log_14_test on node2 Executing query select dictGetInt32('RedisSimple_flat_Int32_', 'Int32_', toUInt64(2)) on node Executing query select dictHas('LocalClickHouse_complex_key_hashed_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select dictHas('MySQL_complex_key_hashed_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query ALTER TABLE tbl ADD COLUMN y Int32 on node1 Executing query drop table if exists system.text_log_15_test on node2 Executing query select dictHas('RedisSimple_flat_Int32_', toUInt64(2)) on node Executing query select dictGetOrDefault('LocalClickHouse_complex_key_hashed_', 'Date_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDate('2018-12-30')) on local_node Executing query select dictGetOrDefault('MySQL_complex_key_hashed_', 'UInt32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt32(77)) on node Executing query INSERT INTO tbl VALUES (2, 20) on node1 Executing query drop table if exists system.query_log_15_test on node2 Executing query select dictGetDateOrDefault('LocalClickHouse_complex_key_hashed_', 'Date_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDate('2018-12-30')) on local_node Executing query drop table if exists system.text_log_16_test on node2 Executing query select dictGetUInt32OrDefault('MySQL_complex_key_hashed_', 'UInt32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt32(77)) on node Executing query select dictGetOrDefault('RedisSimple_flat_Int32_', 'Int32_', toUInt64(9999999), toInt32(-77)) on node Executing query ALTER TABLE tbl ADD COLUMN z Int32 on node1 Executing query drop table if exists system.query_log_16_test on node2 Executing query select dictGet('MySQL_complex_key_hashed_', 'UInt64_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGetInt32OrDefault('RedisSimple_flat_Int32_', 'Int32_', toUInt64(9999999), toInt32(-77)) on node Executing query INSERT INTO tbl VALUES (3, 30, 300) on node1 Executing query select dictGet('LocalClickHouse_complex_key_hashed_', 'DateTime_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query drop table if exists system.text_log_17_test on node2 Executing query select dictGetUInt64('MySQL_complex_key_hashed_', 'UInt64_', tuple(toUInt64(2),toString('qwerty2'))) on node Stderr: Container roottestanalyzercompatibility-gw4-backward-1 Stopping Stderr: Container roottestanalyzercompatibility-gw4-current-1 Stopping Stderr: Container roottestanalyzercompatibility-gw4-backward-1 Stopped Stderr: Container roottestanalyzercompatibility-gw4-current-1 Stopped Command:[bash -c [ -f /ClickHouse/tests/integration/test_analyzer_compatibility/_instances-0-gw4/current/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_analyzer_compatibility/_instances-0-gw4/current/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Executing query BACKUP TABLE tbl ON CLUSTER 'cluster' TO Disk('backups', '18') on node1 Command:[bash -c [ -f /ClickHouse/tests/integration/test_analyzer_compatibility/_instances-0-gw4/backward/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_analyzer_compatibility/_instances-0-gw4/backward/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Executing query select dictGetDateTime('LocalClickHouse_complex_key_hashed_', 'DateTime_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Command:[docker compose --env-file /ClickHouse/tests/integration/test_analyzer_compatibility/_instances-0-gw4/.env --project-name roottestanalyzercompatibility-gw4 --file /ClickHouse/tests/integration/test_analyzer_compatibility/_instances-0-gw4/current/docker-compose.yml --file /ClickHouse/tests/integration/test_analyzer_compatibility/_instances-0-gw4/backward/docker-compose.yml down --volumes] Executing query system reload dictionary RedisSimple_hashed_Int32_ on node Executing query drop table if exists system.query_log_17_test on node2 Executing query select dictGet('RedisSimple_hashed_Int32_', 'Int32_', toUInt64(1)) on node Executing query select dictHas('MySQL_complex_key_hashed_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictHas('LocalClickHouse_complex_key_hashed_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query drop table if exists system.text_log_18_test on node2 Executing query select dictGetInt32('RedisSimple_hashed_Int32_', 'Int32_', toUInt64(1)) on node Executing query select dictGetOrDefault('MySQL_complex_key_hashed_', 'UInt64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt64(88)) on node Executing query select dictGetOrDefault('LocalClickHouse_complex_key_hashed_', 'DateTime_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDateTime('2018-12-30 00:00:00')) on local_node Executing query drop table if exists system.query_log_18_test on node2 Stderr: Container roottestanalyzercompatibility-gw4-backward-1 Stopping Stderr: Container roottestanalyzercompatibility-gw4-current-1 Stopping Stderr: Container roottestanalyzercompatibility-gw4-backward-1 Stopped Stderr: Container roottestanalyzercompatibility-gw4-current-1 Stopped Stderr: Container roottestanalyzercompatibility-gw4-current-1 Removing Stderr: Container roottestanalyzercompatibility-gw4-backward-1 Removing Stderr: Container roottestanalyzercompatibility-gw4-current-1 Removed Stderr: Container roottestanalyzercompatibility-gw4-backward-1 Removed Stderr: Network roottestanalyzercompatibility-gw4_default Removing Stderr: Network roottestanalyzercompatibility-gw4_default Removed Cleanup called Docker networks for project roottestanalyzercompatibility-gw4 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestanalyzercompatibility-gw4 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestanalyzercompatibility-gw4 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestanalyzercompatibility-gw4-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} No running containers for project: roottestanalyzercompatibility-gw4 Trying to prune unused networks... Trying to prune unused images... Command:[docker image prune -f] Executing query DROP TABLE tbl ON CLUSTER 'cluster' SYNC on node1 Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Stdout:3 Command:[docker volume prune -f] Stdout:Total reclaimed space: 0B Volumes pruned: 3 Executing query drop table if exists system.text_log_19_test on node2 Executing query select dictGetUInt64OrDefault('MySQL_complex_key_hashed_', 'UInt64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUInt64(88)) on node Executing query select dictHas('RedisSimple_hashed_Int32_', toUInt64(1)) on node Executing query select dictGetDateTimeOrDefault('LocalClickHouse_complex_key_hashed_', 'DateTime_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDateTime('2018-12-30 00:00:00')) on local_node Executing query drop table if exists system.query_log_19_test on node2 Executing query select dictGet('MySQL_complex_key_hashed_', 'Int8_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGetOrDefault('RedisSimple_hashed_Int32_', 'Int32_', toUInt64(9999999), toInt32(-77)) on node Executing query select dictGet('LocalClickHouse_complex_key_hashed_', 'String_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query RESTORE TABLE tbl ON CLUSTER 'cluster' FROM Disk('backups', '18') on node1 Executing query drop table if exists system.text_log_20_test on node2 Executing query select dictGetInt32OrDefault('RedisSimple_hashed_Int32_', 'Int32_', toUInt64(9999999), toInt32(-77)) on node Executing query select dictGetInt8('MySQL_complex_key_hashed_', 'Int8_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGetString('LocalClickHouse_complex_key_hashed_', 'String_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query drop table if exists system.query_log_20_test on node2 Executing query select dictGet('RedisSimple_hashed_Int32_', 'Int32_', toUInt64(2)) on node Executing query select dictHas('MySQL_complex_key_hashed_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictHas('LocalClickHouse_complex_key_hashed_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query drop table if exists system.text_log_21_test on node2 Executing query OPTIMIZE TABLE tbl FINAL on node1 Executing query select dictGetInt32('RedisSimple_hashed_Int32_', 'Int32_', toUInt64(2)) on node Executing query select dictGetOrDefault('MySQL_complex_key_hashed_', 'Int8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt8(-55)) on node Executing query select dictGetOrDefault('LocalClickHouse_complex_key_hashed_', 'String_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toString('hi')) on local_node Executing query drop table if exists system.query_log_21_test on node2 Executing query SELECT * FROM tbl ORDER BY x on node1 Executing query select dictHas('RedisSimple_hashed_Int32_', toUInt64(2)) on node Executing query select dictGetInt8OrDefault('MySQL_complex_key_hashed_', 'Int8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt8(-55)) on node Executing query drop table if exists system.text_log_22_test on node2 Executing query select dictGetStringOrDefault('LocalClickHouse_complex_key_hashed_', 'String_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toString('hi')) on local_node Executing query SELECT * FROM tbl ORDER BY x on node2 Executing query drop table if exists system.query_log_22_test on node2 Executing query select dictGet('MySQL_complex_key_hashed_', 'Int16_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGetOrDefault('RedisSimple_hashed_Int32_', 'Int32_', toUInt64(9999999), toInt32(-77)) on node Executing query select dictGet('LocalClickHouse_complex_key_hashed_', 'Float32_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query drop table if exists system.text_log_23_test on node2 Executing query DROP TABLE IF EXISTS tbl ON CLUSTER 'cluster3' SYNC on node1 [gw2] PASSED test_backup_restore_on_cluster/test.py::test_replicated_table_after_alters Executing query select dictGetInt16('MySQL_complex_key_hashed_', 'Int16_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGetInt32OrDefault('RedisSimple_hashed_Int32_', 'Int32_', toUInt64(9999999), toInt32(-77)) on node Executing query select dictGetFloat32('LocalClickHouse_complex_key_hashed_', 'Float32_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query drop table if exists system.query_log_23_test on node2 Executing query select dictHas('MySQL_complex_key_hashed_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query DROP TABLE IF EXISTS tbl2 ON CLUSTER 'cluster3' SYNC on node1 Executing query system reload dictionary RedisSimple_cache_Int32_ on node Executing query drop table if exists system.text_log_24_test on node2 Executing query select dictHas('LocalClickHouse_complex_key_hashed_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select dictGetOrDefault('MySQL_complex_key_hashed_', 'Int16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt16(-66)) on node Executing query drop table if exists system.query_log_24_test on node2 Executing query select dictGet('RedisSimple_cache_Int32_', 'Int32_', toUInt64(1)) on node Executing query DROP DATABASE IF EXISTS mydb ON CLUSTER 'cluster3' SYNC on node1 Executing query select dictGetOrDefault('LocalClickHouse_complex_key_hashed_', 'Float32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat32(555.11)) on local_node Executing query drop table if exists system.text_log_25_test on node2 Executing query select dictGetInt16OrDefault('MySQL_complex_key_hashed_', 'Int16_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt16(-66)) on node Executing query select dictGetInt32('RedisSimple_cache_Int32_', 'Int32_', toUInt64(1)) on node Executing query drop table if exists system.query_log_25_test on node2 Executing query select dictGetFloat32OrDefault('LocalClickHouse_complex_key_hashed_', 'Float32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat32(555.11)) on local_node Executing query DROP DATABASE IF EXISTS mydb2 ON CLUSTER 'cluster3' SYNC on node1 Executing query select dictGet('MySQL_complex_key_hashed_', 'Int32_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query drop table if exists system.text_log_26_test on node2 Executing query select dictHas('RedisSimple_cache_Int32_', toUInt64(1)) on node Executing query select dictGet('LocalClickHouse_complex_key_hashed_', 'Float64_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query select dictGetInt32('MySQL_complex_key_hashed_', 'Int32_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query DROP USER IF EXISTS u1, u2 ON CLUSTER 'cluster3' on node1 Executing query drop table if exists system.query_log_26_test on node2 Executing query select dictGetOrDefault('RedisSimple_cache_Int32_', 'Int32_', toUInt64(9999999), toInt32(-77)) on node Executing query select dictHas('MySQL_complex_key_hashed_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGetFloat64('LocalClickHouse_complex_key_hashed_', 'Float64_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query drop table if exists system.text_log_27_test on node2 test_backup_restore_on_cluster/test.py::test_replicated_table_restored_into_bigger_cluster Executing query CREATE TABLE tbl ON CLUSTER 'cluster' (x UInt32) ENGINE=ReplicatedMergeTree('/clickhouse/tables/tbl/', '{replica}')ORDER BY x on node1 Executing query select dictGetInt32OrDefault('RedisSimple_cache_Int32_', 'Int32_', toUInt64(9999999), toInt32(-77)) on node Executing query select dictGetOrDefault('MySQL_complex_key_hashed_', 'Int32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt32(-77)) on node Executing query select dictHas('LocalClickHouse_complex_key_hashed_', tuple(toUInt64(2),toString('qwerty2'))) on local_node Executing query drop table if exists system.query_log_27_test on node2 Executing query select dictGet('RedisSimple_cache_Int32_', 'Int32_', toUInt64(2)) on node Executing query INSERT INTO tbl VALUES (111) on node1 Executing query drop table if exists system.text_log_28_test on node2 Executing query select dictGetInt32OrDefault('MySQL_complex_key_hashed_', 'Int32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt32(-77)) on node Executing query select dictGetOrDefault('LocalClickHouse_complex_key_hashed_', 'Float64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat64(777.11)) on local_node Executing query drop table if exists system.query_log_28_test on node2 Executing query INSERT INTO tbl VALUES (222) on node2 Executing query select dictGetInt32('RedisSimple_cache_Int32_', 'Int32_', toUInt64(2)) on node Executing query select dictGet('MySQL_complex_key_hashed_', 'Int64_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGetFloat64OrDefault('LocalClickHouse_complex_key_hashed_', 'Float64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat64(777.11)) on local_node Executing query drop table if exists system.text_log_29_test on node2 Executing query BACKUP TABLE tbl ON CLUSTER 'cluster' TO Disk('backups', '19') on node1 Executing query select dictHas('RedisSimple_cache_Int32_', toUInt64(2)) on node Executing query select dictGetInt64('MySQL_complex_key_hashed_', 'Int64_', tuple(toUInt64(2),toString('qwerty2'))) on node [gw5] PASSED test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_complex[complex_key_hashed] test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_ranged[range_hashed] Executing query system reload dictionaries on local_node Executing query drop table if exists system.query_log_29_test on node2 Executing query select dictGetOrDefault('RedisSimple_cache_Int32_', 'Int32_', toUInt64(9999999), toInt32(-77)) on node Executing query select dictHas('MySQL_complex_key_hashed_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query drop table if exists system.text_log_30_test on node2 Executing query select dictGet('LocalClickHouse_range_hashed_', 'KeyField2', toUInt64(1), toDate('2019-02-10')) on local_node Executing query select dictGetInt32OrDefault('RedisSimple_cache_Int32_', 'Int32_', toUInt64(9999999), toInt32(-77)) on node Executing query drop table if exists system.query_log_30_test on node2 Executing query select dictGetOrDefault('MySQL_complex_key_hashed_', 'Int64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt64(-88)) on node Executing query DROP TABLE tbl ON CLUSTER 'cluster' SYNC on node1 Executing query select dictGetDate('LocalClickHouse_range_hashed_', 'KeyField2', toUInt64(1), toDate('2019-02-10')) on local_node Executing query drop table if exists system.text_log_31_test on node2 Executing query system reload dictionary RedisSimple_direct_Int32_ on node Executing query select dictGetInt64OrDefault('MySQL_complex_key_hashed_', 'Int64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt64(-88)) on node Executing query select dictGet('LocalClickHouse_range_hashed_', 'UInt8_', toUInt64(1), toDate('2019-02-10')) on local_node Executing query drop table if exists system.query_log_31_test on node2 Executing query select dictGet('RedisSimple_direct_Int32_', 'Int32_', toUInt64(1)) on node Executing query RESTORE TABLE tbl ON CLUSTER 'cluster3' FROM Disk('backups', '19') on node1 Executing query drop table if exists system.text_log_32_test on node2 Executing query select dictGet('MySQL_complex_key_hashed_', 'UUID_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGetUInt8('LocalClickHouse_range_hashed_', 'UInt8_', toUInt64(1), toDate('2019-02-10')) on local_node Executing query select dictGetInt32('RedisSimple_direct_Int32_', 'Int32_', toUInt64(1)) on node Executing query drop table if exists system.query_log_32_test on node2 Executing query select dictGetUUID('MySQL_complex_key_hashed_', 'UUID_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGet('LocalClickHouse_range_hashed_', 'UInt16_', toUInt64(1), toDate('2019-02-10')) on local_node Executing query drop table if exists system.text_log_33_test on node2 Executing query select dictHas('RedisSimple_direct_Int32_', toUInt64(1)) on node Executing query SYSTEM SYNC REPLICA ON CLUSTER 'cluster3' tbl on node1 Executing query select dictHas('MySQL_complex_key_hashed_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGetUInt16('LocalClickHouse_range_hashed_', 'UInt16_', toUInt64(1), toDate('2019-02-10')) on local_node Executing query drop table if exists system.query_log_33_test on node2 Executing query select dictGetOrDefault('RedisSimple_direct_Int32_', 'Int32_', toUInt64(9999999), toInt32(-77)) on node Executing query SELECT * FROM tbl ORDER BY x on node1 Executing query select dictGetOrDefault('MySQL_complex_key_hashed_', 'UUID_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUUID('550e8400-0000-0000-0000-000000000000')) on node Executing query drop table if exists system.text_log_34_test on node2 Executing query select dictGet('LocalClickHouse_range_hashed_', 'UInt32_', toUInt64(1), toDate('2019-02-10')) on local_node Executing query select dictGetInt32OrDefault('RedisSimple_direct_Int32_', 'Int32_', toUInt64(9999999), toInt32(-77)) on node Executing query select dictGetUUIDOrDefault('MySQL_complex_key_hashed_', 'UUID_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toUUID('550e8400-0000-0000-0000-000000000000')) on node Executing query SELECT * FROM tbl ORDER BY x on node2 Executing query drop table if exists system.query_log_34_test on node2 Executing query select dictGetUInt32('LocalClickHouse_range_hashed_', 'UInt32_', toUInt64(1), toDate('2019-02-10')) on local_node Connection dropped: socket connection error: None Executing query select dictGet('RedisSimple_direct_Int32_', 'Int32_', toUInt64(2)) on node Executing query drop table if exists system.text_log_35_test on node2 Executing query select dictGet('MySQL_complex_key_hashed_', 'Date_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGet('LocalClickHouse_range_hashed_', 'UInt64_', toUInt64(1), toDate('2019-02-10')) on local_node Executing query SELECT * FROM tbl ORDER BY x on node3 Executing query drop table if exists system.query_log_35_test on node2 Executing query select dictGetInt32('RedisSimple_direct_Int32_', 'Int32_', toUInt64(2)) on node Executing query select dictGetDate('MySQL_complex_key_hashed_', 'Date_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query drop table if exists system.text_log_36_test on node2 Executing query DROP TABLE IF EXISTS tbl ON CLUSTER 'cluster3' SYNC on node1 [gw2] PASSED test_backup_restore_on_cluster/test.py::test_replicated_table_restored_into_bigger_cluster Executing query select dictGetUInt64('LocalClickHouse_range_hashed_', 'UInt64_', toUInt64(1), toDate('2019-02-10')) on local_node Executing query select dictHas('RedisSimple_direct_Int32_', toUInt64(2)) on node Executing query select dictHas('MySQL_complex_key_hashed_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query drop table if exists system.query_log_36_test on node2 Executing query select dictGet('LocalClickHouse_range_hashed_', 'Int8_', toUInt64(1), toDate('2019-02-10')) on local_node Executing query select dictGetOrDefault('RedisSimple_direct_Int32_', 'Int32_', toUInt64(9999999), toInt32(-77)) on node Executing query select dictGetOrDefault('MySQL_complex_key_hashed_', 'Date_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDate('2018-12-30')) on node Executing query DROP TABLE IF EXISTS tbl2 ON CLUSTER 'cluster3' SYNC on node1 Executing query drop table if exists system.text_log_37_test on node2 Executing query select dictGetInt8('LocalClickHouse_range_hashed_', 'Int8_', toUInt64(1), toDate('2019-02-10')) on local_node Executing query select dictGetInt32OrDefault('RedisSimple_direct_Int32_', 'Int32_', toUInt64(9999999), toInt32(-77)) on node Executing query drop table if exists system.query_log_37_test on node2 Executing query select dictGetDateOrDefault('MySQL_complex_key_hashed_', 'Date_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDate('2018-12-30')) on node Executing query DROP DATABASE IF EXISTS mydb ON CLUSTER 'cluster3' SYNC on node1 Executing query drop table if exists system.text_log_38_test on node2 Executing query select dictGet('LocalClickHouse_range_hashed_', 'Int16_', toUInt64(1), toDate('2019-02-10')) on local_node Executing query system reload dictionary RedisHash_complex_key_hashed_Int32_ on node Executing query select dictGet('MySQL_complex_key_hashed_', 'DateTime_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query drop table if exists system.query_log_38_test on node2 Executing query DROP DATABASE IF EXISTS mydb2 ON CLUSTER 'cluster3' SYNC on node1 Executing query select dictGet('RedisHash_complex_key_hashed_Int32_', 'Int32_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGetInt16('LocalClickHouse_range_hashed_', 'Int16_', toUInt64(1), toDate('2019-02-10')) on local_node Executing query select dictGetDateTime('MySQL_complex_key_hashed_', 'DateTime_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query drop table if exists system.text_log_39_test on node2 Executing query select dictGetInt32('RedisHash_complex_key_hashed_Int32_', 'Int32_', tuple(toUInt64(1),toString('world'))) on node Executing query DROP USER IF EXISTS u1, u2 ON CLUSTER 'cluster3' on node1 Executing query select dictGet('LocalClickHouse_range_hashed_', 'Int32_', toUInt64(1), toDate('2019-02-10')) on local_node Executing query select dictHas('MySQL_complex_key_hashed_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query drop table if exists system.query_log_39_test on node2 Executing query select dictHas('RedisHash_complex_key_hashed_Int32_', tuple(toUInt64(1),toString('world'))) on node Executing query drop table if exists system.text_log_40_test on node2 Executing query select dictGetOrDefault('MySQL_complex_key_hashed_', 'DateTime_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDateTime('2018-12-30 00:00:00')) on node Executing query select dictGetInt32('LocalClickHouse_range_hashed_', 'Int32_', toUInt64(1), toDate('2019-02-10')) on local_node test_backup_restore_on_cluster/test.py::test_replicated_table_restored_into_smaller_cluster Executing query CREATE TABLE tbl ON CLUSTER 'cluster' (x UInt32) ENGINE=ReplicatedMergeTree('/clickhouse/tables/tbl/', '{replica}')ORDER BY x on node1 Executing query select dictGetOrDefault('RedisHash_complex_key_hashed_Int32_', 'Int32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt32(-77)) on node Executing query drop table if exists system.query_log_40_test on node2 Executing query select dictGetDateTimeOrDefault('MySQL_complex_key_hashed_', 'DateTime_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toDateTime('2018-12-30 00:00:00')) on node Executing query select dictGet('LocalClickHouse_range_hashed_', 'Int64_', toUInt64(1), toDate('2019-02-10')) on local_node [gw7] PASSED test_async_load_databases/test.py::test_async_load_system_database test_async_load_databases/test.py::test_dependent_tables Executing query create database lazy engine=Lazy(10) on node1 Executing query select dictGetInt32OrDefault('RedisHash_complex_key_hashed_Int32_', 'Int32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt32(-77)) on node Executing query INSERT INTO tbl VALUES (111) on node1 Executing query select dictGet('MySQL_complex_key_hashed_', 'String_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query create database a on node1 Executing query select dictGetInt64('LocalClickHouse_range_hashed_', 'Int64_', toUInt64(1), toDate('2019-02-10')) on local_node Executing query INSERT INTO tbl VALUES (222) on node2 Executing query select dictGet('RedisHash_complex_key_hashed_Int32_', 'Int32_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query create table lazy.src (n int, m int) engine=Log on node1 Executing query select dictGetString('MySQL_complex_key_hashed_', 'String_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGet('LocalClickHouse_range_hashed_', 'UUID_', toUInt64(1), toDate('2019-02-10')) on local_node Executing query BACKUP TABLE tbl ON CLUSTER 'cluster' TO Disk('backups', '20') on node1 Executing query select dictGetInt32('RedisHash_complex_key_hashed_Int32_', 'Int32_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query create dictionary a.d (n int default 0, m int default 42) primary key n source(clickhouse(host 'localhost' port tcpPort() user 'default' table 'src' password '' db 'lazy'))lifetime(min 1 max 10) layout(flat()) on node1 Executing query select dictGetUUID('LocalClickHouse_range_hashed_', 'UUID_', toUInt64(1), toDate('2019-02-10')) on local_node Executing query select dictHas('MySQL_complex_key_hashed_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query create table system.join (n int, m int) engine=Join(any, left, n) on node1 Executing query select dictHas('RedisHash_complex_key_hashed_Int32_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGet('LocalClickHouse_range_hashed_', 'Date_', toUInt64(1), toDate('2019-02-10')) on local_node Executing query select dictGetOrDefault('MySQL_complex_key_hashed_', 'String_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toString('hi')) on node Executing query insert into system.join values (1, 1) on node1 Executing query select dictGetOrDefault('RedisHash_complex_key_hashed_Int32_', 'Int32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt32(-77)) on node Executing query select dictGetDate('LocalClickHouse_range_hashed_', 'Date_', toUInt64(1), toDate('2019-02-10')) on local_node Executing query select dictGetStringOrDefault('MySQL_complex_key_hashed_', 'String_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toString('hi')) on node Executing query insert into system.join values (1, 2) on node1 Executing query select dictGetInt32OrDefault('RedisHash_complex_key_hashed_Int32_', 'Int32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt32(-77)) on node Executing query DROP TABLE tbl ON CLUSTER 'cluster' SYNC on node1 Executing query select dictGet('MySQL_complex_key_hashed_', 'Float32_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGet('LocalClickHouse_range_hashed_', 'DateTime_', toUInt64(1), toDate('2019-02-10')) on local_node Executing query insert into system.join values (1, 3) on node1 Executing query system reload dictionary RedisHash_complex_key_cache_Int32_ on node Executing query select dictGetFloat32('MySQL_complex_key_hashed_', 'Float32_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGetDateTime('LocalClickHouse_range_hashed_', 'DateTime_', toUInt64(1), toDate('2019-02-10')) on local_node Executing query RESTORE TABLE tbl ON CLUSTER 'cluster1' FROM Disk('backups', '20') on node1 Executing query insert into system.join values (1, 4) on node1 Executing query select dictGet('RedisHash_complex_key_cache_Int32_', 'Int32_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictHas('MySQL_complex_key_hashed_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGet('LocalClickHouse_range_hashed_', 'String_', toUInt64(1), toDate('2019-02-10')) on local_node Executing query select dictGetInt32('RedisHash_complex_key_cache_Int32_', 'Int32_', tuple(toUInt64(1),toString('world'))) on node Executing query insert into system.join values (1, 5) on node1 Executing query select dictGetOrDefault('MySQL_complex_key_hashed_', 'Float32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat32(555.11)) on node Executing query SELECT * FROM tbl ORDER BY x on node1 Executing query select dictGetString('LocalClickHouse_range_hashed_', 'String_', toUInt64(1), toDate('2019-02-10')) on local_node Executing query select dictHas('RedisHash_complex_key_cache_Int32_', tuple(toUInt64(1),toString('world'))) on node Executing query insert into system.join values (1, 6) on node1 Executing query select dictGet('LocalClickHouse_range_hashed_', 'Float32_', toUInt64(1), toDate('2019-02-10')) on local_node Executing query select dictGetFloat32OrDefault('MySQL_complex_key_hashed_', 'Float32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat32(555.11)) on node Executing query DROP TABLE IF EXISTS tbl ON CLUSTER 'cluster3' SYNC on node1 [gw2] PASSED test_backup_restore_on_cluster/test.py::test_replicated_table_restored_into_smaller_cluster Executing query select dictGetOrDefault('RedisHash_complex_key_cache_Int32_', 'Int32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt32(-77)) on node Executing query insert into system.join values (1, 7) on node1 Executing query select dictGet('MySQL_complex_key_hashed_', 'Float64_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGetFloat32('LocalClickHouse_range_hashed_', 'Float32_', toUInt64(1), toDate('2019-02-10')) on local_node Executing query select dictGetInt32OrDefault('RedisHash_complex_key_cache_Int32_', 'Int32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt32(-77)) on node Executing query insert into system.join values (1, 8) on node1 Executing query DROP TABLE IF EXISTS tbl2 ON CLUSTER 'cluster3' SYNC on node1 Executing query select dictGetFloat64('MySQL_complex_key_hashed_', 'Float64_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGet('LocalClickHouse_range_hashed_', 'Float64_', toUInt64(1), toDate('2019-02-10')) on local_node Executing query select dictGet('RedisHash_complex_key_cache_Int32_', 'Int32_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query insert into system.join values (1, 9) on node1 Executing query select dictHas('MySQL_complex_key_hashed_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGetFloat64('LocalClickHouse_range_hashed_', 'Float64_', toUInt64(1), toDate('2019-02-10')) on local_node Executing query DROP DATABASE IF EXISTS mydb ON CLUSTER 'cluster3' SYNC on node1 Executing query select dictGetInt32('RedisHash_complex_key_cache_Int32_', 'Int32_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query insert into system.join values (1, 10) on node1 Executing query select dictGetOrDefault('MySQL_complex_key_hashed_', 'Float64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat64(777.11)) on node Executing query select dictGet('LocalClickHouse_range_hashed_', 'KeyField2', toUInt64(2), toDate('2019-04-10')) on local_node Executing query DROP DATABASE IF EXISTS mydb2 ON CLUSTER 'cluster3' SYNC on node1 Executing query insert into system.join values (1, 11) on node1 Executing query select dictHas('RedisHash_complex_key_cache_Int32_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGetFloat64OrDefault('MySQL_complex_key_hashed_', 'Float64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toFloat64(777.11)) on node Executing query select dictGetDate('LocalClickHouse_range_hashed_', 'KeyField2', toUInt64(2), toDate('2019-04-10')) on local_node Executing query insert into system.join values (1, 12) on node1 Executing query select dictGetOrDefault('RedisHash_complex_key_cache_Int32_', 'Int32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt32(-77)) on node Executing query DROP USER IF EXISTS u1, u2 ON CLUSTER 'cluster3' on node1 [gw3] PASSED test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_complex[complex_key_hashed] test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_ranged[range_hashed] Executing query system reload dictionaries on node Executing query select dictGet('LocalClickHouse_range_hashed_', 'UInt8_', toUInt64(2), toDate('2019-04-10')) on local_node Executing query insert into system.join values (1, 13) on node1 Executing query select dictGetInt32OrDefault('RedisHash_complex_key_cache_Int32_', 'Int32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt32(-77)) on node Executing query select dictGet('MySQL_range_hashed_', 'KeyField2', toUInt64(1), toDate('2019-02-10')) on node Executing query select dictGetUInt8('LocalClickHouse_range_hashed_', 'UInt8_', toUInt64(2), toDate('2019-04-10')) on local_node test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_def Executing query CREATE TABLE tbl (x UInt8, y String) ENGINE=ReplicatedMergeTree('/clickhouse/tables/tbl/', '{replica}')ORDER BY tuple() on node1 Executing query insert into system.join values (1, 14) on node1 Executing query system reload dictionary RedisHash_complex_key_direct_Int32_ on node Executing query select dictGet('LocalClickHouse_range_hashed_', 'UInt16_', toUInt64(2), toDate('2019-04-10')) on local_node Executing query select dictGetDate('MySQL_range_hashed_', 'KeyField2', toUInt64(1), toDate('2019-02-10')) on node Executing query CREATE TABLE tbl (x UInt8, y String) ENGINE=ReplicatedMergeTree('/clickhouse/tables/tbl/', '{replica}')ORDER BY tuple() on node2 Executing query select dictGet('RedisHash_complex_key_direct_Int32_', 'Int32_', tuple(toUInt64(1),toString('world'))) on node Executing query insert into system.join values (1, 15) on node1 Executing query select dictGetUInt16('LocalClickHouse_range_hashed_', 'UInt16_', toUInt64(2), toDate('2019-04-10')) on local_node Executing query select dictGet('MySQL_range_hashed_', 'UInt8_', toUInt64(1), toDate('2019-02-10')) on node Executing query SYSTEM STOP REPLICATION QUEUES tbl on node2 Executing query select dictGetInt32('RedisHash_complex_key_direct_Int32_', 'Int32_', tuple(toUInt64(1),toString('world'))) on node Executing query insert into system.join values (1, 16) on node1 Executing query select dictGet('LocalClickHouse_range_hashed_', 'UInt32_', toUInt64(2), toDate('2019-04-10')) on local_node Executing query select dictGetUInt8('MySQL_range_hashed_', 'UInt8_', toUInt64(1), toDate('2019-02-10')) on node Executing query ALTER TABLE tbl MODIFY COLUMN x String on node1 Executing query select dictHas('RedisHash_complex_key_direct_Int32_', tuple(toUInt64(1),toString('world'))) on node Executing query insert into system.join values (1, 17) on node1 Executing query select dictGetUInt32('LocalClickHouse_range_hashed_', 'UInt32_', toUInt64(2), toDate('2019-04-10')) on local_node Executing query select dictGet('MySQL_range_hashed_', 'UInt16_', toUInt64(1), toDate('2019-02-10')) on node Executing query SELECT name, type FROM system.columns WHERE database='default' AND table='tbl' on node1 Executing query select dictGetOrDefault('RedisHash_complex_key_direct_Int32_', 'Int32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt32(-77)) on node Executing query insert into system.join values (1, 18) on node1 Executing query select dictGet('LocalClickHouse_range_hashed_', 'UInt64_', toUInt64(2), toDate('2019-04-10')) on local_node Executing query select dictGetUInt16('MySQL_range_hashed_', 'UInt16_', toUInt64(1), toDate('2019-02-10')) on node Executing query SELECT name, type FROM system.columns WHERE database='default' AND table='tbl' on node2 Executing query select dictGetInt32OrDefault('RedisHash_complex_key_direct_Int32_', 'Int32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt32(-77)) on node Executing query insert into system.join values (1, 19) on node1 Executing query select dictGet('MySQL_range_hashed_', 'UInt32_', toUInt64(1), toDate('2019-02-10')) on node Executing query BACKUP TABLE tbl ON CLUSTER 'cluster' TO Disk('backups', '21') on node2 Executing query select dictGetUInt64('LocalClickHouse_range_hashed_', 'UInt64_', toUInt64(2), toDate('2019-04-10')) on local_node Executing query insert into system.join values (1, 20) on node1 Executing query select dictGet('RedisHash_complex_key_direct_Int32_', 'Int32_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGetUInt32('MySQL_range_hashed_', 'UInt32_', toUInt64(1), toDate('2019-02-10')) on node Executing query select dictGet('LocalClickHouse_range_hashed_', 'Int8_', toUInt64(2), toDate('2019-04-10')) on local_node Executing query insert into system.join values (1, 21) on node1 Executing query select dictGetInt32('RedisHash_complex_key_direct_Int32_', 'Int32_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGetInt8('LocalClickHouse_range_hashed_', 'Int8_', toUInt64(2), toDate('2019-04-10')) on local_node Executing query select dictGet('MySQL_range_hashed_', 'UInt64_', toUInt64(1), toDate('2019-02-10')) on node Executing query DROP TABLE tbl ON CLUSTER 'cluster' SYNC on node1 Executing query insert into system.join values (1, 22) on node1 Executing query select dictHas('RedisHash_complex_key_direct_Int32_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGet('LocalClickHouse_range_hashed_', 'Int16_', toUInt64(2), toDate('2019-04-10')) on local_node Executing query select dictGetUInt64('MySQL_range_hashed_', 'UInt64_', toUInt64(1), toDate('2019-02-10')) on node Executing query RESTORE TABLE tbl ON CLUSTER 'cluster' FROM Disk('backups', '21') SETTINGS replica_num_in_backup=1 on node1 Executing query insert into system.join values (1, 23) on node1 Executing query select dictGetOrDefault('RedisHash_complex_key_direct_Int32_', 'Int32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt32(-77)) on node Executing query select dictGetInt16('LocalClickHouse_range_hashed_', 'Int16_', toUInt64(2), toDate('2019-04-10')) on local_node Executing query select dictGet('MySQL_range_hashed_', 'Int8_', toUInt64(1), toDate('2019-02-10')) on node Executing query select dictGetInt32OrDefault('RedisHash_complex_key_direct_Int32_', 'Int32_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt32(-77)) on node Executing query insert into system.join values (1, 24) on node1 Executing query select dictGet('LocalClickHouse_range_hashed_', 'Int32_', toUInt64(2), toDate('2019-04-10')) on local_node Executing query select dictGetInt8('MySQL_range_hashed_', 'Int8_', toUInt64(1), toDate('2019-02-10')) on node Executing query SELECT name, type FROM system.columns WHERE database='default' AND table='tbl' on node1 Executing query system reload dictionaries on node Executing query insert into system.join values (1, 25) on node1 Executing query select dictGetInt32('LocalClickHouse_range_hashed_', 'Int32_', toUInt64(2), toDate('2019-04-10')) on local_node Executing query select dictGet('MySQL_range_hashed_', 'Int16_', toUInt64(1), toDate('2019-02-10')) on node [gw0] PASSED test_dictionaries_redis/test.py::test_redis_dictionaries[Int32_] test_dictionaries_redis/test.py::test_redis_dictionaries[Int64_] Run test with id: 7 Executing query system reload dictionaries on node Executing query insert into system.join values (1, 26) on node1 Executing query SELECT name, type FROM system.columns WHERE database='default' AND table='tbl' on node2 Executing query select dictGet('LocalClickHouse_range_hashed_', 'Int64_', toUInt64(2), toDate('2019-04-10')) on local_node Executing query select dictGetInt16('MySQL_range_hashed_', 'Int16_', toUInt64(1), toDate('2019-02-10')) on node Executing query system reload dictionary RedisSimple_flat_Int64_ on node Executing query insert into system.join values (1, 27) on node1 Executing query DROP TABLE tbl ON CLUSTER 'cluster' SYNC on node1 Executing query select dictGetInt64('LocalClickHouse_range_hashed_', 'Int64_', toUInt64(2), toDate('2019-04-10')) on local_node Executing query select dictGet('MySQL_range_hashed_', 'Int32_', toUInt64(1), toDate('2019-02-10')) on node Executing query select dictGet('RedisSimple_flat_Int64_', 'Int64_', toUInt64(1)) on node Executing query insert into system.join values (1, 28) on node1 Executing query select dictGet('LocalClickHouse_range_hashed_', 'UUID_', toUInt64(2), toDate('2019-04-10')) on local_node Executing query RESTORE TABLE tbl ON CLUSTER 'cluster' FROM Disk('backups', '21') SETTINGS replica_num_in_backup=2 on node2 Executing query select dictGetInt32('MySQL_range_hashed_', 'Int32_', toUInt64(1), toDate('2019-02-10')) on node Executing query select dictGetInt64('RedisSimple_flat_Int64_', 'Int64_', toUInt64(1)) on node Executing query insert into system.join values (1, 29) on node1 Executing query select dictGetUUID('LocalClickHouse_range_hashed_', 'UUID_', toUInt64(2), toDate('2019-04-10')) on local_node Executing query select dictGet('MySQL_range_hashed_', 'Int64_', toUInt64(1), toDate('2019-02-10')) on node Executing query select dictHas('RedisSimple_flat_Int64_', toUInt64(1)) on node Executing query insert into system.join values (1, 30) on node1 Executing query select dictGet('LocalClickHouse_range_hashed_', 'Date_', toUInt64(2), toDate('2019-04-10')) on local_node Executing query select dictGetInt64('MySQL_range_hashed_', 'Int64_', toUInt64(1), toDate('2019-02-10')) on node Executing query select dictGetOrDefault('RedisSimple_flat_Int64_', 'Int64_', toUInt64(9999999), toInt64(-88)) on node Executing query SELECT name, type FROM system.columns WHERE database='default' AND table='tbl' on node1 Executing query insert into system.join values (1, 31) on node1 Executing query select dictGetDate('LocalClickHouse_range_hashed_', 'Date_', toUInt64(2), toDate('2019-04-10')) on local_node Executing query select dictGet('MySQL_range_hashed_', 'UUID_', toUInt64(1), toDate('2019-02-10')) on node Executing query select dictGetInt64OrDefault('RedisSimple_flat_Int64_', 'Int64_', toUInt64(9999999), toInt64(-88)) on node Executing query SELECT name, type FROM system.columns WHERE database='default' AND table='tbl' on node2 Executing query insert into system.join values (1, 32) on node1 Executing query select dictGet('LocalClickHouse_range_hashed_', 'DateTime_', toUInt64(2), toDate('2019-04-10')) on local_node Executing query select dictGetUUID('MySQL_range_hashed_', 'UUID_', toUInt64(1), toDate('2019-02-10')) on node Executing query select dictGet('RedisSimple_flat_Int64_', 'Int64_', toUInt64(2)) on node Executing query DROP TABLE IF EXISTS tbl ON CLUSTER 'cluster3' SYNC on node1 [gw2] PASSED test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_def Executing query insert into system.join values (1, 33) on node1 Executing query select dictGet('MySQL_range_hashed_', 'Date_', toUInt64(1), toDate('2019-02-10')) on node Executing query select dictGetInt64('RedisSimple_flat_Int64_', 'Int64_', toUInt64(2)) on node Executing query select dictGetDateTime('LocalClickHouse_range_hashed_', 'DateTime_', toUInt64(2), toDate('2019-04-10')) on local_node Executing query insert into system.join values (1, 34) on node1 Executing query DROP TABLE IF EXISTS tbl2 ON CLUSTER 'cluster3' SYNC on node1 Executing query select dictGetDate('MySQL_range_hashed_', 'Date_', toUInt64(1), toDate('2019-02-10')) on node Executing query select dictHas('RedisSimple_flat_Int64_', toUInt64(2)) on node Executing query select dictGet('LocalClickHouse_range_hashed_', 'String_', toUInt64(2), toDate('2019-04-10')) on local_node Executing query insert into system.join values (1, 35) on node1 Executing query select dictGet('MySQL_range_hashed_', 'DateTime_', toUInt64(1), toDate('2019-02-10')) on node Executing query DROP DATABASE IF EXISTS mydb ON CLUSTER 'cluster3' SYNC on node1 Executing query select dictGetString('LocalClickHouse_range_hashed_', 'String_', toUInt64(2), toDate('2019-04-10')) on local_node Executing query select dictGetOrDefault('RedisSimple_flat_Int64_', 'Int64_', toUInt64(9999999), toInt64(-88)) on node Executing query insert into system.join values (1, 36) on node1 Executing query select dictGetDateTime('MySQL_range_hashed_', 'DateTime_', toUInt64(1), toDate('2019-02-10')) on node Executing query select dictGet('LocalClickHouse_range_hashed_', 'Float32_', toUInt64(2), toDate('2019-04-10')) on local_node Executing query select dictGetInt64OrDefault('RedisSimple_flat_Int64_', 'Int64_', toUInt64(9999999), toInt64(-88)) on node Executing query insert into system.join values (1, 37) on node1 Executing query DROP DATABASE IF EXISTS mydb2 ON CLUSTER 'cluster3' SYNC on node1 Executing query select dictGet('MySQL_range_hashed_', 'String_', toUInt64(1), toDate('2019-02-10')) on node Executing query select dictGetFloat32('LocalClickHouse_range_hashed_', 'Float32_', toUInt64(2), toDate('2019-04-10')) on local_node Executing query system reload dictionary RedisSimple_hashed_Int64_ on node Executing query insert into system.join values (1, 38) on node1 Executing query DROP USER IF EXISTS u1, u2 ON CLUSTER 'cluster3' on node1 Executing query select dictGetString('MySQL_range_hashed_', 'String_', toUInt64(1), toDate('2019-02-10')) on node Executing query select dictGet('LocalClickHouse_range_hashed_', 'Float64_', toUInt64(2), toDate('2019-04-10')) on local_node Executing query select dictGet('RedisSimple_hashed_Int64_', 'Int64_', toUInt64(1)) on node Executing query insert into system.join values (1, 39) on node1 Executing query select dictGet('MySQL_range_hashed_', 'Float32_', toUInt64(1), toDate('2019-02-10')) on node test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_insert Executing query CREATE TABLE tbl ON CLUSTER 'cluster' (x UInt32) ENGINE=ReplicatedMergeTree('/clickhouse/tables/tbl/', '{replica}')ORDER BY x on node1 Executing query select dictGetFloat64('LocalClickHouse_range_hashed_', 'Float64_', toUInt64(2), toDate('2019-04-10')) on local_node Executing query select dictGetInt64('RedisSimple_hashed_Int64_', 'Int64_', toUInt64(1)) on node Executing query insert into system.join values (1, 40) on node1 Executing query select dictGetFloat32('MySQL_range_hashed_', 'Float32_', toUInt64(1), toDate('2019-02-10')) on node [gw5] PASSED test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_ranged[range_hashed] test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_simple[cache] Executing query system reload dictionaries on local_node Executing query select dictHas('RedisSimple_hashed_Int64_', toUInt64(1)) on node Executing query insert into system.join values (1, 41) on node1 Executing query INSERT INTO tbl VALUES (111) on node1 Executing query select dictGet('MySQL_range_hashed_', 'Float64_', toUInt64(1), toDate('2019-02-10')) on node Executing query select dictGet('LocalClickHouse_cache_', 'UInt8_', toUInt64(1)) on local_node Executing query insert into system.join values (1, 42) on node1 Executing query select dictGetOrDefault('RedisSimple_hashed_Int64_', 'Int64_', toUInt64(9999999), toInt64(-88)) on node Executing query INSERT INTO tbl VALUES (222) on node2 Executing query select dictGetFloat64('MySQL_range_hashed_', 'Float64_', toUInt64(1), toDate('2019-02-10')) on node Executing query select dictGetUInt8('LocalClickHouse_cache_', 'UInt8_', toUInt64(1)) on local_node Executing query select dictGetInt64OrDefault('RedisSimple_hashed_Int64_', 'Int64_', toUInt64(9999999), toInt64(-88)) on node Executing query insert into system.join values (1, 43) on node1 Executing query SYSTEM SYNC REPLICA ON CLUSTER 'cluster' tbl on node1 Executing query select dictGet('MySQL_range_hashed_', 'KeyField2', toUInt64(2), toDate('2019-04-10')) on node Executing query select dictHas('LocalClickHouse_cache_', toUInt64(1)) on local_node Executing query select dictGet('RedisSimple_hashed_Int64_', 'Int64_', toUInt64(2)) on node Executing query insert into system.join values (1, 44) on node1 Executing query SYSTEM STOP REPLICATED SENDS ON CLUSTER 'cluster' tbl on node1 Executing query select dictGetDate('MySQL_range_hashed_', 'KeyField2', toUInt64(2), toDate('2019-04-10')) on node Executing query select dictGetOrDefault('LocalClickHouse_cache_', 'UInt8_', toUInt64(9999999), toUInt8(55)) on local_node Executing query select dictGetInt64('RedisSimple_hashed_Int64_', 'Int64_', toUInt64(2)) on node Executing query insert into system.join values (1, 45) on node1 Executing query select dictGet('MySQL_range_hashed_', 'UInt8_', toUInt64(2), toDate('2019-04-10')) on node Executing query select dictHas('RedisSimple_hashed_Int64_', toUInt64(2)) on node Executing query INSERT INTO tbl VALUES (333) on node1 Executing query select dictGetUInt8OrDefault('LocalClickHouse_cache_', 'UInt8_', toUInt64(9999999), toUInt8(55)) on local_node Executing query insert into system.join values (1, 46) on node1 Executing query select dictGetUInt8('MySQL_range_hashed_', 'UInt8_', toUInt64(2), toDate('2019-04-10')) on node Executing query INSERT INTO tbl VALUES (444) on node2 Executing query select dictGet('LocalClickHouse_cache_', 'UInt16_', toUInt64(1)) on local_node Executing query select dictGetOrDefault('RedisSimple_hashed_Int64_', 'Int64_', toUInt64(9999999), toInt64(-88)) on node Executing query insert into system.join values (1, 47) on node1 Executing query select dictGet('MySQL_range_hashed_', 'UInt16_', toUInt64(2), toDate('2019-04-10')) on node Executing query BACKUP TABLE tbl ON CLUSTER 'cluster' TO Disk('backups', '22') on node1 Executing query select dictGetInt64OrDefault('RedisSimple_hashed_Int64_', 'Int64_', toUInt64(9999999), toInt64(-88)) on node Executing query insert into system.join values (1, 48) on node1 Executing query select dictGetUInt16('LocalClickHouse_cache_', 'UInt16_', toUInt64(1)) on local_node Executing query select dictGetUInt16('MySQL_range_hashed_', 'UInt16_', toUInt64(2), toDate('2019-04-10')) on node Executing query system reload dictionary RedisSimple_cache_Int64_ on node Executing query insert into system.join values (1, 49) on node1 Executing query select dictHas('LocalClickHouse_cache_', toUInt64(1)) on local_node Executing query select dictGet('RedisSimple_cache_Int64_', 'Int64_', toUInt64(1)) on node Executing query select dictGet('MySQL_range_hashed_', 'UInt32_', toUInt64(2), toDate('2019-04-10')) on node Executing query insert into system.join values (1, 50) on node1 Executing query DROP TABLE tbl ON CLUSTER 'cluster' SYNC on node1 Executing query select dictGetOrDefault('LocalClickHouse_cache_', 'UInt16_', toUInt64(9999999), toUInt16(66)) on local_node Executing query select dictGetInt64('RedisSimple_cache_Int64_', 'Int64_', toUInt64(1)) on node Executing query select dictGetUInt32('MySQL_range_hashed_', 'UInt32_', toUInt64(2), toDate('2019-04-10')) on node Executing query insert into system.join values (1, 51) on node1 Executing query select dictGetUInt16OrDefault('LocalClickHouse_cache_', 'UInt16_', toUInt64(9999999), toUInt16(66)) on local_node Executing query RESTORE TABLE tbl ON CLUSTER 'cluster' FROM Disk('backups', '22') on node1 Executing query select dictGet('MySQL_range_hashed_', 'UInt64_', toUInt64(2), toDate('2019-04-10')) on node Executing query select dictHas('RedisSimple_cache_Int64_', toUInt64(1)) on node Executing query insert into system.join values (1, 52) on node1 Executing query select dictGet('LocalClickHouse_cache_', 'UInt32_', toUInt64(1)) on local_node Executing query select dictGetUInt64('MySQL_range_hashed_', 'UInt64_', toUInt64(2), toDate('2019-04-10')) on node Executing query select dictGetOrDefault('RedisSimple_cache_Int64_', 'Int64_', toUInt64(9999999), toInt64(-88)) on node Executing query insert into system.join values (1, 53) on node1 Executing query select dictGetUInt32('LocalClickHouse_cache_', 'UInt32_', toUInt64(1)) on local_node Executing query select dictGet('MySQL_range_hashed_', 'Int8_', toUInt64(2), toDate('2019-04-10')) on node Executing query select dictGetInt64OrDefault('RedisSimple_cache_Int64_', 'Int64_', toUInt64(9999999), toInt64(-88)) on node Executing query insert into system.join values (1, 54) on node1 Executing query SYSTEM SYNC REPLICA ON CLUSTER 'cluster' tbl on node1 Executing query select dictHas('LocalClickHouse_cache_', toUInt64(1)) on local_node Executing query select dictGetInt8('MySQL_range_hashed_', 'Int8_', toUInt64(2), toDate('2019-04-10')) on node Executing query select dictGet('RedisSimple_cache_Int64_', 'Int64_', toUInt64(2)) on node Executing query insert into system.join values (1, 55) on node1 Executing query SELECT * FROM tbl ORDER BY x on node1 Executing query select dictGetOrDefault('LocalClickHouse_cache_', 'UInt32_', toUInt64(9999999), toUInt32(77)) on local_node Executing query select dictGet('MySQL_range_hashed_', 'Int16_', toUInt64(2), toDate('2019-04-10')) on node Executing query select dictGetInt64('RedisSimple_cache_Int64_', 'Int64_', toUInt64(2)) on node Executing query insert into system.join values (1, 56) on node1 Executing query SELECT * FROM tbl ORDER BY x on node2 Executing query select dictGetUInt32OrDefault('LocalClickHouse_cache_', 'UInt32_', toUInt64(9999999), toUInt32(77)) on local_node Executing query select dictGetInt16('MySQL_range_hashed_', 'Int16_', toUInt64(2), toDate('2019-04-10')) on node Executing query select dictHas('RedisSimple_cache_Int64_', toUInt64(2)) on node Executing query insert into system.join values (1, 57) on node1 Executing query DROP TABLE IF EXISTS tbl ON CLUSTER 'cluster3' SYNC on node1 [gw2] PASSED test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_insert Executing query select dictGet('LocalClickHouse_cache_', 'UInt64_', toUInt64(1)) on local_node Executing query select dictGet('MySQL_range_hashed_', 'Int32_', toUInt64(2), toDate('2019-04-10')) on node Executing query select dictGetOrDefault('RedisSimple_cache_Int64_', 'Int64_', toUInt64(9999999), toInt64(-88)) on node Executing query insert into system.join values (1, 58) on node1 Executing query DROP TABLE IF EXISTS tbl2 ON CLUSTER 'cluster3' SYNC on node1 Executing query select dictGetUInt64('LocalClickHouse_cache_', 'UInt64_', toUInt64(1)) on local_node Executing query select dictGetInt32('MySQL_range_hashed_', 'Int32_', toUInt64(2), toDate('2019-04-10')) on node Executing query select dictGetInt64OrDefault('RedisSimple_cache_Int64_', 'Int64_', toUInt64(9999999), toInt64(-88)) on node Executing query insert into system.join values (1, 59) on node1 Executing query select dictHas('LocalClickHouse_cache_', toUInt64(1)) on local_node Executing query select dictGet('MySQL_range_hashed_', 'Int64_', toUInt64(2), toDate('2019-04-10')) on node Executing query DROP DATABASE IF EXISTS mydb ON CLUSTER 'cluster3' SYNC on node1 Executing query system reload dictionary RedisSimple_direct_Int64_ on node Executing query insert into system.join values (1, 60) on node1 Executing query select dictGetOrDefault('LocalClickHouse_cache_', 'UInt64_', toUInt64(9999999), toUInt64(88)) on local_node Executing query select dictGetInt64('MySQL_range_hashed_', 'Int64_', toUInt64(2), toDate('2019-04-10')) on node Executing query select dictGet('RedisSimple_direct_Int64_', 'Int64_', toUInt64(1)) on node Executing query DROP DATABASE IF EXISTS mydb2 ON CLUSTER 'cluster3' SYNC on node1 Executing query insert into system.join values (1, 61) on node1 Executing query select dictGetUInt64OrDefault('LocalClickHouse_cache_', 'UInt64_', toUInt64(9999999), toUInt64(88)) on local_node Executing query select dictGetInt64('RedisSimple_direct_Int64_', 'Int64_', toUInt64(1)) on node Executing query select dictGet('MySQL_range_hashed_', 'UUID_', toUInt64(2), toDate('2019-04-10')) on node Executing query DROP USER IF EXISTS u1, u2 ON CLUSTER 'cluster3' on node1 Executing query insert into system.join values (1, 62) on node1 Executing query select dictHas('RedisSimple_direct_Int64_', toUInt64(1)) on node Executing query select dictGet('LocalClickHouse_cache_', 'Int8_', toUInt64(1)) on local_node Executing query select dictGetUUID('MySQL_range_hashed_', 'UUID_', toUInt64(2), toDate('2019-04-10')) on node Executing query insert into system.join values (1, 63) on node1 test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_merge Executing query CREATE TABLE tbl ON CLUSTER 'cluster' (x UInt32) ENGINE=ReplicatedMergeTree('/clickhouse/tables/tbl/', '{replica}')ORDER BY x on node1 Executing query select dictGetInt8('LocalClickHouse_cache_', 'Int8_', toUInt64(1)) on local_node Executing query select dictGet('MySQL_range_hashed_', 'Date_', toUInt64(2), toDate('2019-04-10')) on node Executing query select dictGetOrDefault('RedisSimple_direct_Int64_', 'Int64_', toUInt64(9999999), toInt64(-88)) on node Connection dropped: socket connection error: None Executing query insert into system.join values (1, 64) on node1 Executing query select dictGetInt64OrDefault('RedisSimple_direct_Int64_', 'Int64_', toUInt64(9999999), toInt64(-88)) on node Executing query select dictHas('LocalClickHouse_cache_', toUInt64(1)) on local_node Executing query select dictGetDate('MySQL_range_hashed_', 'Date_', toUInt64(2), toDate('2019-04-10')) on node Executing query SYSTEM STOP MERGES ON CLUSTER 'cluster' tbl on node1 Executing query insert into system.join values (1, 65) on node1 Executing query select dictGet('MySQL_range_hashed_', 'DateTime_', toUInt64(2), toDate('2019-04-10')) on node Executing query select dictGet('RedisSimple_direct_Int64_', 'Int64_', toUInt64(2)) on node Executing query select dictGetOrDefault('LocalClickHouse_cache_', 'Int8_', toUInt64(9999999), toInt8(-55)) on local_node Executing query INSERT INTO tbl VALUES (111) on node1 Executing query insert into system.join values (1, 66) on node1 Executing query select dictGetDateTime('MySQL_range_hashed_', 'DateTime_', toUInt64(2), toDate('2019-04-10')) on node Executing query select dictGetInt8OrDefault('LocalClickHouse_cache_', 'Int8_', toUInt64(9999999), toInt8(-55)) on local_node Executing query select dictGetInt64('RedisSimple_direct_Int64_', 'Int64_', toUInt64(2)) on node Executing query INSERT INTO tbl VALUES (222) on node1 Executing query insert into system.join values (1, 67) on node1 Executing query select dictGet('MySQL_range_hashed_', 'String_', toUInt64(2), toDate('2019-04-10')) on node Executing query select dictHas('RedisSimple_direct_Int64_', toUInt64(2)) on node Executing query select dictGet('LocalClickHouse_cache_', 'Int16_', toUInt64(1)) on local_node Executing query SYSTEM SYNC REPLICA tbl on node2 Executing query insert into system.join values (1, 68) on node1 Executing query select dictGetString('MySQL_range_hashed_', 'String_', toUInt64(2), toDate('2019-04-10')) on node Executing query select dictGetOrDefault('RedisSimple_direct_Int64_', 'Int64_', toUInt64(9999999), toInt64(-88)) on node Executing query select dictGetInt16('LocalClickHouse_cache_', 'Int16_', toUInt64(1)) on local_node Executing query SYSTEM START MERGES tbl on node2 Executing query insert into system.join values (1, 69) on node1 Executing query select dictGet('MySQL_range_hashed_', 'Float32_', toUInt64(2), toDate('2019-04-10')) on node Executing query OPTIMIZE TABLE tbl FINAL on node2 Executing query select dictGetInt64OrDefault('RedisSimple_direct_Int64_', 'Int64_', toUInt64(9999999), toInt64(-88)) on node Executing query select dictHas('LocalClickHouse_cache_', toUInt64(1)) on local_node Executing query insert into system.join values (1, 70) on node1 Executing query select dictGetFloat32('MySQL_range_hashed_', 'Float32_', toUInt64(2), toDate('2019-04-10')) on node Executing query BACKUP TABLE tbl ON CLUSTER 'cluster' TO Disk('backups', '23') on node1 Executing query system reload dictionary RedisHash_complex_key_hashed_Int64_ on node Executing query select dictGetOrDefault('LocalClickHouse_cache_', 'Int16_', toUInt64(9999999), toInt16(-66)) on local_node Executing query insert into system.join values (1, 71) on node1 Executing query select dictGet('MySQL_range_hashed_', 'Float64_', toUInt64(2), toDate('2019-04-10')) on node Executing query select dictGet('RedisHash_complex_key_hashed_Int64_', 'Int64_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGetInt16OrDefault('LocalClickHouse_cache_', 'Int16_', toUInt64(9999999), toInt16(-66)) on local_node Executing query insert into system.join values (1, 72) on node1 Executing query select dictGetFloat64('MySQL_range_hashed_', 'Float64_', toUInt64(2), toDate('2019-04-10')) on node Executing query select dictGetInt64('RedisHash_complex_key_hashed_Int64_', 'Int64_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGet('LocalClickHouse_cache_', 'Int32_', toUInt64(1)) on local_node Executing query DROP TABLE tbl ON CLUSTER 'cluster' SYNC on node1 Executing query insert into system.join values (1, 73) on node1 [gw3] PASSED test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_ranged[range_hashed] test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_simple[cache] Executing query system reload dictionaries on node Executing query select dictHas('RedisHash_complex_key_hashed_Int64_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGetInt32('LocalClickHouse_cache_', 'Int32_', toUInt64(1)) on local_node Executing query insert into system.join values (1, 74) on node1 Executing query RESTORE TABLE tbl ON CLUSTER 'cluster' FROM Disk('backups', '23') on node1 Executing query select dictGet('MySQL_cache_', 'UInt8_', toUInt64(1)) on node Executing query select dictGetOrDefault('RedisHash_complex_key_hashed_Int64_', 'Int64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt64(-88)) on node Executing query select dictHas('LocalClickHouse_cache_', toUInt64(1)) on local_node Executing query insert into system.join values (1, 75) on node1 Executing query select dictGetUInt8('MySQL_cache_', 'UInt8_', toUInt64(1)) on node Executing query select dictGetInt64OrDefault('RedisHash_complex_key_hashed_Int64_', 'Int64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt64(-88)) on node Executing query select dictGetOrDefault('LocalClickHouse_cache_', 'Int32_', toUInt64(9999999), toInt32(-77)) on local_node Executing query insert into system.join values (1, 76) on node1 Executing query select dictHas('MySQL_cache_', toUInt64(1)) on node Executing query SYSTEM SYNC REPLICA ON CLUSTER 'cluster' tbl on node1 Executing query select dictGet('RedisHash_complex_key_hashed_Int64_', 'Int64_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGetInt32OrDefault('LocalClickHouse_cache_', 'Int32_', toUInt64(9999999), toInt32(-77)) on local_node Executing query insert into system.join values (1, 77) on node1 Executing query select dictGetOrDefault('MySQL_cache_', 'UInt8_', toUInt64(9999999), toUInt8(55)) on node Executing query select dictGetInt64('RedisHash_complex_key_hashed_Int64_', 'Int64_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGet('LocalClickHouse_cache_', 'Int64_', toUInt64(1)) on local_node Executing query SELECT * FROM tbl ORDER BY x on node1 Executing query insert into system.join values (1, 78) on node1 Executing query select dictGetUInt8OrDefault('MySQL_cache_', 'UInt8_', toUInt64(9999999), toUInt8(55)) on node Executing query select dictHas('RedisHash_complex_key_hashed_Int64_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGetInt64('LocalClickHouse_cache_', 'Int64_', toUInt64(1)) on local_node Executing query SELECT * FROM tbl ORDER BY x on node2 Executing query insert into system.join values (1, 79) on node1 Executing query select dictGet('MySQL_cache_', 'UInt16_', toUInt64(1)) on node Executing query select dictGetOrDefault('RedisHash_complex_key_hashed_Int64_', 'Int64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt64(-88)) on node Executing query select dictHas('LocalClickHouse_cache_', toUInt64(1)) on local_node Executing query DROP TABLE IF EXISTS tbl ON CLUSTER 'cluster3' SYNC on node1 [gw2] PASSED test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_merge Executing query insert into system.join values (1, 80) on node1 Executing query select dictGetUInt16('MySQL_cache_', 'UInt16_', toUInt64(1)) on node Executing query select dictGetInt64OrDefault('RedisHash_complex_key_hashed_Int64_', 'Int64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt64(-88)) on node Executing query select dictGetOrDefault('LocalClickHouse_cache_', 'Int64_', toUInt64(9999999), toInt64(-88)) on local_node Executing query insert into system.join values (1, 81) on node1 Executing query DROP TABLE IF EXISTS tbl2 ON CLUSTER 'cluster3' SYNC on node1 Executing query select dictHas('MySQL_cache_', toUInt64(1)) on node Executing query system reload dictionary RedisHash_complex_key_cache_Int64_ on node Executing query select dictGetInt64OrDefault('LocalClickHouse_cache_', 'Int64_', toUInt64(9999999), toInt64(-88)) on local_node Executing query insert into system.join values (1, 82) on node1 Executing query select dictGetOrDefault('MySQL_cache_', 'UInt16_', toUInt64(9999999), toUInt16(66)) on node Executing query DROP DATABASE IF EXISTS mydb ON CLUSTER 'cluster3' SYNC on node1 Executing query select dictGet('RedisHash_complex_key_cache_Int64_', 'Int64_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGet('LocalClickHouse_cache_', 'UUID_', toUInt64(1)) on local_node Executing query insert into system.join values (1, 83) on node1 Executing query select dictGetUInt16OrDefault('MySQL_cache_', 'UInt16_', toUInt64(9999999), toUInt16(66)) on node Executing query select dictGetInt64('RedisHash_complex_key_cache_Int64_', 'Int64_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGetUUID('LocalClickHouse_cache_', 'UUID_', toUInt64(1)) on local_node Executing query DROP DATABASE IF EXISTS mydb2 ON CLUSTER 'cluster3' SYNC on node1 Executing query insert into system.join values (1, 84) on node1 Executing query select dictGet('MySQL_cache_', 'UInt32_', toUInt64(1)) on node Executing query select dictHas('RedisHash_complex_key_cache_Int64_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictHas('LocalClickHouse_cache_', toUInt64(1)) on local_node Executing query insert into system.join values (1, 85) on node1 Executing query DROP USER IF EXISTS u1, u2 ON CLUSTER 'cluster3' on node1 Executing query select dictGetUInt32('MySQL_cache_', 'UInt32_', toUInt64(1)) on node Executing query select dictGetOrDefault('RedisHash_complex_key_cache_Int64_', 'Int64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt64(-88)) on node Executing query select dictGetOrDefault('LocalClickHouse_cache_', 'UUID_', toUInt64(9999999), toUUID('550e8400-0000-0000-0000-000000000000')) on local_node Executing query insert into system.join values (1, 86) on node1 Executing query select dictHas('MySQL_cache_', toUInt64(1)) on node test_backup_restore_on_cluster/test.py::test_replicated_table_with_uuid_in_zkpath Executing query CREATE TABLE tbl ON CLUSTER 'cluster' (x UInt8, y String) ENGINE=ReplicatedMergeTree('/clickhouse/tables/{uuid}','{replica}')ORDER BY x on node1 Executing query select dictGetInt64OrDefault('RedisHash_complex_key_cache_Int64_', 'Int64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt64(-88)) on node Executing query insert into system.join values (1, 87) on node1 Executing query select dictGetUUIDOrDefault('LocalClickHouse_cache_', 'UUID_', toUInt64(9999999), toUUID('550e8400-0000-0000-0000-000000000000')) on local_node Executing query select dictGetOrDefault('MySQL_cache_', 'UInt32_', toUInt64(9999999), toUInt32(77)) on node Executing query select dictGet('RedisHash_complex_key_cache_Int64_', 'Int64_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query insert into system.join values (1, 88) on node1 Executing query select dictGet('LocalClickHouse_cache_', 'Date_', toUInt64(1)) on local_node Executing query INSERT INTO tbl VALUES (1, 'AA') on node1 Executing query select dictGetUInt32OrDefault('MySQL_cache_', 'UInt32_', toUInt64(9999999), toUInt32(77)) on node Executing query select dictGetInt64('RedisHash_complex_key_cache_Int64_', 'Int64_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query insert into system.join values (1, 89) on node1 Executing query select dictGetDate('LocalClickHouse_cache_', 'Date_', toUInt64(1)) on local_node Executing query INSERT INTO tbl VALUES (2, 'BB') on node2 Executing query select dictGet('MySQL_cache_', 'UInt64_', toUInt64(1)) on node Executing query select dictHas('RedisHash_complex_key_cache_Int64_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query insert into system.join values (1, 90) on node1 Executing query select dictHas('LocalClickHouse_cache_', toUInt64(1)) on local_node Executing query BACKUP TABLE tbl ON CLUSTER 'cluster' TO Disk('backups', '24') on node1 Executing query select dictGetUInt64('MySQL_cache_', 'UInt64_', toUInt64(1)) on node Executing query select dictGetOrDefault('RedisHash_complex_key_cache_Int64_', 'Int64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt64(-88)) on node Executing query select dictGetOrDefault('LocalClickHouse_cache_', 'Date_', toUInt64(9999999), toDate('2018-12-30')) on local_node Executing query insert into system.join values (1, 91) on node1 Executing query select dictHas('MySQL_cache_', toUInt64(1)) on node Executing query select dictGetInt64OrDefault('RedisHash_complex_key_cache_Int64_', 'Int64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt64(-88)) on node Executing query insert into system.join values (1, 92) on node1 Executing query select dictGetDateOrDefault('LocalClickHouse_cache_', 'Date_', toUInt64(9999999), toDate('2018-12-30')) on local_node Executing query RESTORE TABLE tbl AS tbl2 ON CLUSTER 'cluster' FROM Disk('backups', '24') on node2 Executing query select dictGetOrDefault('MySQL_cache_', 'UInt64_', toUInt64(9999999), toUInt64(88)) on node Executing query system reload dictionary RedisHash_complex_key_direct_Int64_ on node Executing query insert into system.join values (1, 93) on node1 Executing query select dictGet('LocalClickHouse_cache_', 'DateTime_', toUInt64(1)) on local_node Executing query select dictGetUInt64OrDefault('MySQL_cache_', 'UInt64_', toUInt64(9999999), toUInt64(88)) on node Executing query select dictGet('RedisHash_complex_key_direct_Int64_', 'Int64_', tuple(toUInt64(1),toString('world'))) on node Executing query insert into system.join values (1, 94) on node1 Executing query select dictGetDateTime('LocalClickHouse_cache_', 'DateTime_', toUInt64(1)) on local_node Executing query select dictGet('MySQL_cache_', 'Int8_', toUInt64(1)) on node Executing query INSERT INTO tbl2 VALUES (3, 'CC') on node1 Executing query select dictGetInt64('RedisHash_complex_key_direct_Int64_', 'Int64_', tuple(toUInt64(1),toString('world'))) on node Executing query insert into system.join values (1, 95) on node1 Executing query select dictHas('LocalClickHouse_cache_', toUInt64(1)) on local_node Executing query SYSTEM SYNC REPLICA ON CLUSTER 'cluster' tbl on node1 Executing query select dictGetInt8('MySQL_cache_', 'Int8_', toUInt64(1)) on node Executing query select dictHas('RedisHash_complex_key_direct_Int64_', tuple(toUInt64(1),toString('world'))) on node Executing query insert into system.join values (1, 96) on node1 Executing query select dictGetOrDefault('LocalClickHouse_cache_', 'DateTime_', toUInt64(9999999), toDateTime('2018-12-30 00:00:00')) on local_node Executing query select dictHas('MySQL_cache_', toUInt64(1)) on node Executing query SYSTEM SYNC REPLICA ON CLUSTER 'cluster' tbl2 on node1 Executing query select dictGetOrDefault('RedisHash_complex_key_direct_Int64_', 'Int64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt64(-88)) on node Executing query insert into system.join values (1, 97) on node1 Executing query select dictGetDateTimeOrDefault('LocalClickHouse_cache_', 'DateTime_', toUInt64(9999999), toDateTime('2018-12-30 00:00:00')) on local_node Executing query select dictGetOrDefault('MySQL_cache_', 'Int8_', toUInt64(9999999), toInt8(-55)) on node Executing query select dictGetInt64OrDefault('RedisHash_complex_key_direct_Int64_', 'Int64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt64(-88)) on node Executing query insert into system.join values (1, 98) on node1 Executing query select dictGet('LocalClickHouse_cache_', 'String_', toUInt64(1)) on local_node Executing query SELECT * FROM tbl ORDER BY x on node1 Executing query select dictGetInt8OrDefault('MySQL_cache_', 'Int8_', toUInt64(9999999), toInt8(-55)) on node Executing query select dictGet('RedisHash_complex_key_direct_Int64_', 'Int64_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query insert into system.join values (1, 99) on node1 Executing query select dictGetString('LocalClickHouse_cache_', 'String_', toUInt64(1)) on local_node Executing query SELECT * FROM tbl2 ORDER BY x on node1 Executing query select dictGet('MySQL_cache_', 'Int16_', toUInt64(1)) on node Executing query select dictGetInt64('RedisHash_complex_key_direct_Int64_', 'Int64_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query create table src (n int, m default joinGet('system.join', 'm', 1::int),t default dictGetOrNull('a.d', 'm', toUInt64(3)),k default dictGet('a.d', 'm', toUInt64(4))) engine=MergeTree order by n on node1 Executing query select dictHas('LocalClickHouse_cache_', toUInt64(1)) on local_node Executing query SELECT * FROM tbl ORDER BY x on node2 Executing query select dictGetInt16('MySQL_cache_', 'Int16_', toUInt64(1)) on node Executing query select dictHas('RedisHash_complex_key_direct_Int64_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query create dictionary test.d (n int default 0, m int default 42) primary key n source(clickhouse(host 'localhost' port tcpPort() user 'default' table 'src' password '' db 'default'))lifetime(min 1 max 10) layout(flat()) on node1 Executing query select dictGetOrDefault('LocalClickHouse_cache_', 'String_', toUInt64(9999999), toString('hi')) on local_node Executing query SELECT * FROM tbl2 ORDER BY x on node2 Executing query create table join (n int, m default dictGet('a.d', 'm', toUInt64(3)),k default dictGet('test.d', 'm', toUInt64(0))) engine=Join(any, left, n) on node1 Executing query select dictHas('MySQL_cache_', toUInt64(1)) on node Executing query select dictGetOrDefault('RedisHash_complex_key_direct_Int64_', 'Int64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt64(-88)) on node Executing query select dictGetStringOrDefault('LocalClickHouse_cache_', 'String_', toUInt64(9999999), toString('hi')) on local_node Executing query DROP TABLE IF EXISTS tbl ON CLUSTER 'cluster3' SYNC on node1 [gw2] PASSED test_backup_restore_on_cluster/test.py::test_replicated_table_with_uuid_in_zkpath Executing query create table lazy.log (n default dictGet(test.d, 'm', toUInt64(0))) engine=Log on node1 Executing query select dictGetOrDefault('MySQL_cache_', 'Int16_', toUInt64(9999999), toInt16(-66)) on node Executing query select dictGetInt64OrDefault('RedisHash_complex_key_direct_Int64_', 'Int64_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt64(-88)) on node Executing query create table a.t (n default joinGet('system.join', 'm', 1::int),m default dictGet('test.d', 'm', toUInt64(3)),k default joinGet(join, 'm', 1::int)) engine=MergeTree order by n on node1 Executing query select dictGet('LocalClickHouse_cache_', 'Float32_', toUInt64(1)) on local_node Executing query select dictGetInt16OrDefault('MySQL_cache_', 'Int16_', toUInt64(9999999), toInt16(-66)) on node Executing query system reload dictionaries on node Executing query DROP TABLE IF EXISTS tbl2 ON CLUSTER 'cluster3' SYNC on node1 Executing query select database || '.' || name from system.tables on node1 Executing query select dictGetFloat32('LocalClickHouse_cache_', 'Float32_', toUInt64(1)) on local_node [gw0] PASSED test_dictionaries_redis/test.py::test_redis_dictionaries[Int64_] test_dictionaries_redis/test.py::test_redis_dictionaries[Int8_] Run test with id: 4 Executing query system reload dictionaries on node Executing query select dictGet('MySQL_cache_', 'Int32_', toUInt64(1)) on node run container_id:roottestasyncloaddatabases-gw7-node1-1 detach:False nothrow:True cmd: ['bash', '-c', 'ps -C clickhouse'] Command:[docker exec -u root roottestasyncloaddatabases-gw7-node1-1 bash -c ps -C clickhouse] Stdout: PID TTY TIME CMD Stdout: 8 ? 00:00:37 clickhouse run container_id:roottestasyncloaddatabases-gw7-node1-1 detach:False nothrow:False cmd: ['bash', '-c', 'pkill clickhouse'] Command:[docker exec -u root roottestasyncloaddatabases-gw7-node1-1 bash -c pkill clickhouse] run container_id:roottestasyncloaddatabases-gw7-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw7-node1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query DROP DATABASE IF EXISTS mydb ON CLUSTER 'cluster3' SYNC on node1 Executing query system reload dictionary RedisSimple_flat_Int8_ on node Stdout:8 Executing query select dictHas('LocalClickHouse_cache_', toUInt64(1)) on local_node Executing query select dictGetInt32('MySQL_cache_', 'Int32_', toUInt64(1)) on node Executing query select dictGet('RedisSimple_flat_Int8_', 'Int8_', toUInt64(1)) on node Connection dropped: socket connection error: None Executing query select dictGetOrDefault('LocalClickHouse_cache_', 'Float32_', toUInt64(9999999), toFloat32(555.11)) on local_node Executing query select dictHas('MySQL_cache_', toUInt64(1)) on node Executing query DROP DATABASE IF EXISTS mydb2 ON CLUSTER 'cluster3' SYNC on node1 Executing query select dictGetInt8('RedisSimple_flat_Int8_', 'Int8_', toUInt64(1)) on node Executing query select dictGetOrDefault('MySQL_cache_', 'Int32_', toUInt64(9999999), toInt32(-77)) on node Executing query select dictGetFloat32OrDefault('LocalClickHouse_cache_', 'Float32_', toUInt64(9999999), toFloat32(555.11)) on local_node Executing query DROP USER IF EXISTS u1, u2 ON CLUSTER 'cluster3' on node1 Executing query select dictHas('RedisSimple_flat_Int8_', toUInt64(1)) on node Executing query select dictGet('LocalClickHouse_cache_', 'Float64_', toUInt64(1)) on local_node Executing query select dictGetInt32OrDefault('MySQL_cache_', 'Int32_', toUInt64(9999999), toInt32(-77)) on node run container_id:roottestasyncloaddatabases-gw7-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw7-node1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select dictGetOrDefault('RedisSimple_flat_Int8_', 'Int8_', toUInt64(9999999), toInt8(-55)) on node Stdout:8 test_backup_restore_on_cluster/test.py::test_required_privileges Executing query CREATE TABLE tbl ON CLUSTER 'cluster' (x UInt8) ENGINE=ReplicatedMergeTree('/clickhouse/tables/tbl/', '{replica}')ORDER BY x on node1 Executing query select dictGet('MySQL_cache_', 'Int64_', toUInt64(1)) on node Executing query select dictGetFloat64('LocalClickHouse_cache_', 'Float64_', toUInt64(1)) on local_node Executing query select dictGetInt8OrDefault('RedisSimple_flat_Int8_', 'Int8_', toUInt64(9999999), toInt8(-55)) on node Executing query select dictHas('LocalClickHouse_cache_', toUInt64(1)) on local_node Executing query select dictGetInt64('MySQL_cache_', 'Int64_', toUInt64(1)) on node Executing query INSERT INTO tbl VALUES (100) on node1 Executing query select dictGet('RedisSimple_flat_Int8_', 'Int8_', toUInt64(2)) on node Executing query select dictGetOrDefault('LocalClickHouse_cache_', 'Float64_', toUInt64(9999999), toFloat64(777.11)) on local_node Executing query select dictHas('MySQL_cache_', toUInt64(1)) on node Executing query CREATE USER u1 on node1 Executing query select dictGetInt8('RedisSimple_flat_Int8_', 'Int8_', toUInt64(2)) on node Executing query select dictGetFloat64OrDefault('LocalClickHouse_cache_', 'Float64_', toUInt64(9999999), toFloat64(777.11)) on local_node Executing query select dictGetOrDefault('MySQL_cache_', 'Int64_', toUInt64(9999999), toInt64(-88)) on node Executing query GRANT CLUSTER ON *.* TO u1 on node1 run container_id:roottestasyncloaddatabases-gw7-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw7-node1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:8 Executing query select dictHas('RedisSimple_flat_Int8_', toUInt64(2)) on node Executing query BACKUP TABLE tbl ON CLUSTER 'cluster' TO Disk('backups', '25') on node1 Executing query select dictGetInt64OrDefault('MySQL_cache_', 'Int64_', toUInt64(9999999), toInt64(-88)) on node Executing query select dictGet('LocalClickHouse_cache_', 'ParentKeyField', toUInt64(1)) on local_node Executing query select dictGetOrDefault('RedisSimple_flat_Int8_', 'Int8_', toUInt64(9999999), toInt8(-55)) on node Executing query select dictGet('MySQL_cache_', 'UUID_', toUInt64(1)) on node Executing query select dictGetUInt64('LocalClickHouse_cache_', 'ParentKeyField', toUInt64(1)) on local_node Executing query GRANT BACKUP ON tbl TO u1 on node1 Executing query select dictGetInt8OrDefault('RedisSimple_flat_Int8_', 'Int8_', toUInt64(9999999), toInt8(-55)) on node Executing query select dictGetUUID('MySQL_cache_', 'UUID_', toUInt64(1)) on node Executing query select dictHas('LocalClickHouse_cache_', toUInt64(1)) on local_node Executing query BACKUP TABLE tbl ON CLUSTER 'cluster' TO Disk('backups', '25') on node1 Executing query system reload dictionary RedisSimple_hashed_Int8_ on node Executing query select dictHas('MySQL_cache_', toUInt64(1)) on node run container_id:roottestasyncloaddatabases-gw7-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw7-node1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select dictGetOrDefault('LocalClickHouse_cache_', 'ParentKeyField', toUInt64(9999999), toUInt64(444)) on local_node Stdout:8 Executing query select dictGet('RedisSimple_hashed_Int8_', 'Int8_', toUInt64(1)) on node Executing query select dictGetOrDefault('MySQL_cache_', 'UUID_', toUInt64(9999999), toUUID('550e8400-0000-0000-0000-000000000000')) on node Executing query select dictGetUInt64OrDefault('LocalClickHouse_cache_', 'ParentKeyField', toUInt64(9999999), toUInt64(444)) on local_node Executing query select dictGetInt8('RedisSimple_hashed_Int8_', 'Int8_', toUInt64(1)) on node Executing query DROP TABLE tbl ON CLUSTER 'cluster' SYNC on node1 Executing query select dictGetUUIDOrDefault('MySQL_cache_', 'UUID_', toUInt64(9999999), toUUID('550e8400-0000-0000-0000-000000000000')) on node Executing query select dictGet('LocalClickHouse_cache_', 'UInt8_', toUInt64(2)) on local_node Executing query select dictHas('RedisSimple_hashed_Int8_', toUInt64(1)) on node Executing query select dictGet('MySQL_cache_', 'Date_', toUInt64(1)) on node Executing query RESTORE TABLE tbl AS tbl2 ON CLUSTER 'cluster' FROM Disk('backups', '25') on node1 Executing query select dictGetUInt8('LocalClickHouse_cache_', 'UInt8_', toUInt64(2)) on local_node Executing query select dictGetOrDefault('RedisSimple_hashed_Int8_', 'Int8_', toUInt64(9999999), toInt8(-55)) on node run container_id:roottestasyncloaddatabases-gw7-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw7-node1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select dictGetDate('MySQL_cache_', 'Date_', toUInt64(1)) on node Stdout:8 Executing query select dictHas('LocalClickHouse_cache_', toUInt64(2)) on local_node Executing query GRANT INSERT, CREATE TABLE ON tbl2 TO u1 on node1 Executing query select dictGetInt8OrDefault('RedisSimple_hashed_Int8_', 'Int8_', toUInt64(9999999), toInt8(-55)) on node Executing query select dictHas('MySQL_cache_', toUInt64(1)) on node Executing query RESTORE TABLE tbl AS tbl2 ON CLUSTER 'cluster' FROM Disk('backups', '25') on node1 Executing query select dictGetOrDefault('LocalClickHouse_cache_', 'UInt8_', toUInt64(9999999), toUInt8(55)) on local_node Executing query select dictGet('RedisSimple_hashed_Int8_', 'Int8_', toUInt64(2)) on node Executing query select dictGetOrDefault('MySQL_cache_', 'Date_', toUInt64(9999999), toDate('2018-12-30')) on node Executing query select dictGetUInt8OrDefault('LocalClickHouse_cache_', 'UInt8_', toUInt64(9999999), toUInt8(55)) on local_node Executing query select dictGetInt8('RedisSimple_hashed_Int8_', 'Int8_', toUInt64(2)) on node Executing query select dictGetDateOrDefault('MySQL_cache_', 'Date_', toUInt64(9999999), toDate('2018-12-30')) on node Executing query select dictGet('LocalClickHouse_cache_', 'UInt16_', toUInt64(2)) on local_node run container_id:roottestasyncloaddatabases-gw7-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw7-node1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select dictHas('RedisSimple_hashed_Int8_', toUInt64(2)) on node run container_id:roottestasyncloaddatabases-gw7-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw7-node1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select dictGet('MySQL_cache_', 'DateTime_', toUInt64(1)) on node No clickhouse process running. Start new one. http://localhost:None "POST /v1.46/containers/roottestasyncloaddatabases-gw7-node1-1/exec HTTP/1.1" 201 74 http://localhost:None "POST /v1.46/exec/023836a4ae98432cd6686acb77a7aa338cd37190a9819c983894ce02919449b8/start HTTP/1.1" 200 0 http://localhost:None "GET /v1.46/exec/023836a4ae98432cd6686acb77a7aa338cd37190a9819c983894ce02919449b8/json HTTP/1.1" 200 586 Executing query SYSTEM SYNC REPLICA ON CLUSTER 'cluster' tbl2 on node2 Executing query select dictGetUInt16('LocalClickHouse_cache_', 'UInt16_', toUInt64(2)) on local_node Executing query select dictGetOrDefault('RedisSimple_hashed_Int8_', 'Int8_', toUInt64(9999999), toInt8(-55)) on node Executing query select dictGetDateTime('MySQL_cache_', 'DateTime_', toUInt64(1)) on node Executing query SELECT * FROM tbl2 on node2 Executing query select dictHas('LocalClickHouse_cache_', toUInt64(2)) on local_node Executing query select dictGetInt8OrDefault('RedisSimple_hashed_Int8_', 'Int8_', toUInt64(9999999), toInt8(-55)) on node Executing query select dictHas('MySQL_cache_', toUInt64(1)) on node Executing query DROP TABLE tbl2 ON CLUSTER 'cluster' SYNC on node1 Executing query select dictGetOrDefault('LocalClickHouse_cache_', 'UInt16_', toUInt64(9999999), toUInt16(66)) on local_node Executing query select dictGetOrDefault('MySQL_cache_', 'DateTime_', toUInt64(9999999), toDateTime('2018-12-30 00:00:00')) on node Executing query system reload dictionary RedisSimple_cache_Int8_ on node run container_id:roottestasyncloaddatabases-gw7-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw7-node1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:774 Clickhouse process running. run container_id:roottestasyncloaddatabases-gw7-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw7-node1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select dictGetUInt16OrDefault('LocalClickHouse_cache_', 'UInt16_', toUInt64(9999999), toUInt16(66)) on local_node Stdout:774 Executing query select 20 on node1 Executing query select dictGet('RedisSimple_cache_Int8_', 'Int8_', toUInt64(1)) on node Executing query REVOKE ALL FROM u1 on node1 Executing query select dictGetDateTimeOrDefault('MySQL_cache_', 'DateTime_', toUInt64(9999999), toDateTime('2018-12-30 00:00:00')) on node Executing query select dictGetInt8('RedisSimple_cache_Int8_', 'Int8_', toUInt64(1)) on node Executing query RESTORE ALL ON CLUSTER 'cluster' FROM Disk('backups', '25') on node1 Executing query select dictGet('LocalClickHouse_cache_', 'UInt32_', toUInt64(2)) on local_node Executing query select dictGet('MySQL_cache_', 'String_', toUInt64(1)) on node Executing query GRANT INSERT, CREATE TABLE ON tbl TO u1 on node1 Executing query select dictGetUInt32('LocalClickHouse_cache_', 'UInt32_', toUInt64(2)) on local_node Executing query select dictHas('RedisSimple_cache_Int8_', toUInt64(1)) on node Executing query select dictGetString('MySQL_cache_', 'String_', toUInt64(1)) on node Executing query select 20 on node1 Executing query RESTORE ALL ON CLUSTER 'cluster' FROM Disk('backups', '25') on node1 Executing query select dictHas('LocalClickHouse_cache_', toUInt64(2)) on local_node Executing query select dictGetOrDefault('RedisSimple_cache_Int8_', 'Int8_', toUInt64(9999999), toInt8(-55)) on node Executing query select dictHas('MySQL_cache_', toUInt64(1)) on node Executing query select dictGetOrDefault('LocalClickHouse_cache_', 'UInt32_', toUInt64(9999999), toUInt32(77)) on local_node Executing query select dictGetInt8OrDefault('RedisSimple_cache_Int8_', 'Int8_', toUInt64(9999999), toInt8(-55)) on node Executing query select dictGetOrDefault('MySQL_cache_', 'String_', toUInt64(9999999), toString('hi')) on node Executing query select 20 on node1 Executing query select dictGetStringOrDefault('MySQL_cache_', 'String_', toUInt64(9999999), toString('hi')) on node Executing query SYSTEM SYNC REPLICA ON CLUSTER 'cluster' tbl on node2 Executing query select dictGetUInt32OrDefault('LocalClickHouse_cache_', 'UInt32_', toUInt64(9999999), toUInt32(77)) on local_node Executing query select dictGet('RedisSimple_cache_Int8_', 'Int8_', toUInt64(2)) on node Executing query select dictGetInt8('RedisSimple_cache_Int8_', 'Int8_', toUInt64(2)) on node Executing query select dictGet('MySQL_cache_', 'Float32_', toUInt64(1)) on node Executing query select dictGet('LocalClickHouse_cache_', 'UInt64_', toUInt64(2)) on local_node Executing query SELECT * FROM tbl on node2 Executing query select dictHas('RedisSimple_cache_Int8_', toUInt64(2)) on node Executing query select 20 on node1 Executing query select dictGetUInt64('LocalClickHouse_cache_', 'UInt64_', toUInt64(2)) on local_node Executing query DROP TABLE IF EXISTS tbl ON CLUSTER 'cluster3' SYNC on node1 [gw2] PASSED test_backup_restore_on_cluster/test.py::test_required_privileges Executing query select dictGetFloat32('MySQL_cache_', 'Float32_', toUInt64(1)) on node Executing query select dictHas('LocalClickHouse_cache_', toUInt64(2)) on local_node Executing query select dictGetOrDefault('RedisSimple_cache_Int8_', 'Int8_', toUInt64(9999999), toInt8(-55)) on node Executing query select dictHas('MySQL_cache_', toUInt64(1)) on node Executing query DROP TABLE IF EXISTS tbl2 ON CLUSTER 'cluster3' SYNC on node1 Executing query select dictGetInt8OrDefault('RedisSimple_cache_Int8_', 'Int8_', toUInt64(9999999), toInt8(-55)) on node Executing query select dictGetOrDefault('LocalClickHouse_cache_', 'UInt64_', toUInt64(9999999), toUInt64(88)) on local_node Executing query select 20 on node1 Executing query select dictGetOrDefault('MySQL_cache_', 'Float32_', toUInt64(9999999), toFloat32(555.11)) on node Executing query system reload dictionary RedisSimple_direct_Int8_ on node Executing query DROP DATABASE IF EXISTS mydb ON CLUSTER 'cluster3' SYNC on node1 Executing query select dictGetUInt64OrDefault('LocalClickHouse_cache_', 'UInt64_', toUInt64(9999999), toUInt64(88)) on local_node Executing query select database || '.' || name from system.tables on node1 Executing query select dictGetFloat32OrDefault('MySQL_cache_', 'Float32_', toUInt64(9999999), toFloat32(555.11)) on node Executing query select dictGet('RedisSimple_direct_Int8_', 'Int8_', toUInt64(1)) on node Executing query select dictGet('LocalClickHouse_cache_', 'Int8_', toUInt64(2)) on local_node Executing query select dictGet('MySQL_cache_', 'Float64_', toUInt64(1)) on node Executing query DROP DATABASE IF EXISTS mydb2 ON CLUSTER 'cluster3' SYNC on node1 Executing query drop table a.t on node1 Executing query select dictGetInt8('RedisSimple_direct_Int8_', 'Int8_', toUInt64(1)) on node Executing query select dictGetInt8('LocalClickHouse_cache_', 'Int8_', toUInt64(2)) on local_node Executing query select dictGetFloat64('MySQL_cache_', 'Float64_', toUInt64(1)) on node Executing query drop table lazy.log on node1 Executing query select dictHas('RedisSimple_direct_Int8_', toUInt64(1)) on node Executing query DROP USER IF EXISTS u1, u2 ON CLUSTER 'cluster3' on node1 Executing query select dictHas('LocalClickHouse_cache_', toUInt64(2)) on local_node Executing query select dictHas('MySQL_cache_', toUInt64(1)) on node Executing query drop table join on node1 Executing query select dictGetOrDefault('RedisSimple_direct_Int8_', 'Int8_', toUInt64(9999999), toInt8(-55)) on node test_backup_restore_on_cluster/test.py::test_shutdown_waits_for_backup Executing query CREATE TABLE tbl ON CLUSTER 'cluster' (x UInt8) ENGINE=ReplicatedMergeTree('/clickhouse/tables/tbl/', '{replica}')ORDER BY x on node1 Executing query drop dictionary test.d on node1 Executing query select dictGetOrDefault('LocalClickHouse_cache_', 'Int8_', toUInt64(9999999), toInt8(-55)) on local_node Executing query select dictGetOrDefault('MySQL_cache_', 'Float64_', toUInt64(9999999), toFloat64(777.11)) on node Executing query select dictGetInt8OrDefault('RedisSimple_direct_Int8_', 'Int8_', toUInt64(9999999), toInt8(-55)) on node Executing query drop table src on node1 Executing query select dictGetInt8OrDefault('LocalClickHouse_cache_', 'Int8_', toUInt64(9999999), toInt8(-55)) on local_node Executing query select dictGetFloat64OrDefault('MySQL_cache_', 'Float64_', toUInt64(9999999), toFloat64(777.11)) on node Executing query INSERT INTO tbl VALUES (3) on node1 Executing query select dictGet('RedisSimple_direct_Int8_', 'Int8_', toUInt64(2)) on node Executing query drop table system.join on node1 Executing query select dictGet('MySQL_cache_', 'ParentKeyField', toUInt64(1)) on node Executing query select dictGet('LocalClickHouse_cache_', 'Int16_', toUInt64(2)) on local_node Executing query INSERT INTO tbl VALUES (5) on node2 Executing query drop database a on node1 Executing query select dictGetInt8('RedisSimple_direct_Int8_', 'Int8_', toUInt64(2)) on node Executing query select dictGetUInt64('MySQL_cache_', 'ParentKeyField', toUInt64(1)) on node Executing query select dictGetInt16('LocalClickHouse_cache_', 'Int16_', toUInt64(2)) on local_node Executing query drop database lazy on node1 Executing query BACKUP TABLE tbl ON CLUSTER 'cluster' TO Disk('backups', '26') ASYNC on node1 Executing query select dictHas('RedisSimple_direct_Int8_', toUInt64(2)) on node Executing query select dictHas('MySQL_cache_', toUInt64(1)) on node [gw7] PASSED test_async_load_databases/test.py::test_dependent_tables test_async_load_databases/test.py::test_dict_get_data Executing query CREATE TABLE test.elements (id UInt64, a String, b Int32, c Float64) ENGINE=Log; on node1 Executing query select dictHas('LocalClickHouse_cache_', toUInt64(2)) on local_node run container_id:roottestbackuprestoreoncluster-gw2-node2-1 detach:False nothrow:True cmd: ['bash', '-c', 'ps -C clickhouse'] Command:[docker exec -u root roottestbackuprestoreoncluster-gw2-node2-1 bash -c ps -C clickhouse] Stdout: PID TTY TIME CMD Stdout: 8 ? 00:00:56 clickhouse run container_id:roottestbackuprestoreoncluster-gw2-node2-1 detach:False nothrow:False cmd: ['bash', '-c', 'pkill clickhouse'] Command:[docker exec -u root roottestbackuprestoreoncluster-gw2-node2-1 bash -c pkill clickhouse] Executing query select dictGetOrDefault('RedisSimple_direct_Int8_', 'Int8_', toUInt64(9999999), toInt8(-55)) on node run container_id:roottestbackuprestoreoncluster-gw2-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackuprestoreoncluster-gw2-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query INSERT INTO test.elements VALUES (0, 'water', 10, 1), (1, 'air', 40, 0.01), (2, 'earth', 100, 1.7); on node1 Stdout:8 Executing query select dictGetOrDefault('MySQL_cache_', 'ParentKeyField', toUInt64(9999999), toUInt64(444)) on node Executing query select dictGetOrDefault('LocalClickHouse_cache_', 'Int16_', toUInt64(9999999), toInt16(-66)) on local_node Executing query select dictGetInt8OrDefault('RedisSimple_direct_Int8_', 'Int8_', toUInt64(9999999), toInt8(-55)) on node Executing query select dictGetUInt64OrDefault('MySQL_cache_', 'ParentKeyField', toUInt64(9999999), toUInt64(444)) on node Executing query SELECT status FROM system.dictionaries WHERE name='dep_x' on node1 Executing query select dictGetInt16OrDefault('LocalClickHouse_cache_', 'Int16_', toUInt64(9999999), toInt16(-66)) on local_node Executing query system reload dictionary RedisHash_complex_key_hashed_Int8_ on node Executing query SELECT status FROM system.dictionaries WHERE name='dep_y' on node1 Executing query select dictGet('MySQL_cache_', 'UInt8_', toUInt64(2)) on node Executing query select dictGet('LocalClickHouse_cache_', 'Int32_', toUInt64(2)) on local_node Executing query select dictGet('RedisHash_complex_key_hashed_Int8_', 'Int8_', tuple(toUInt64(1),toString('world'))) on node Executing query SELECT status FROM system.dictionaries WHERE name='dep_z' on node1 Executing query select dictGetUInt8('MySQL_cache_', 'UInt8_', toUInt64(2)) on node Executing query select dictGetInt32('LocalClickHouse_cache_', 'Int32_', toUInt64(2)) on local_node run container_id:roottestbackuprestoreoncluster-gw2-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackuprestoreoncluster-gw2-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:8 Executing query select dictGetInt8('RedisHash_complex_key_hashed_Int8_', 'Int8_', tuple(toUInt64(1),toString('world'))) on node Executing query SELECT dictGetString('dep_x', 'a', toUInt64(1)) on node1 Executing query select dictHas('MySQL_cache_', toUInt64(2)) on node Executing query select dictHas('LocalClickHouse_cache_', toUInt64(2)) on local_node Executing query select dictHas('RedisHash_complex_key_hashed_Int8_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGetOrDefault('MySQL_cache_', 'UInt8_', toUInt64(9999999), toUInt8(55)) on node Executing query SELECT status FROM system.dictionaries WHERE name='dep_x' on node1 Executing query select dictGetOrDefault('LocalClickHouse_cache_', 'Int32_', toUInt64(9999999), toInt32(-77)) on local_node Executing query select dictGetOrDefault('RedisHash_complex_key_hashed_Int8_', 'Int8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt8(-55)) on node Executing query select dictGetUInt8OrDefault('MySQL_cache_', 'UInt8_', toUInt64(9999999), toUInt8(55)) on node Executing query SELECT status FROM system.dictionaries WHERE name='dep_y' on node1 Executing query select dictGetInt32OrDefault('LocalClickHouse_cache_', 'Int32_', toUInt64(9999999), toInt32(-77)) on local_node Executing query select dictGetInt8OrDefault('RedisHash_complex_key_hashed_Int8_', 'Int8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt8(-55)) on node Executing query SELECT status FROM system.dictionaries WHERE name='dep_z' on node1 Executing query select dictGet('MySQL_cache_', 'UInt16_', toUInt64(2)) on node run container_id:roottestbackuprestoreoncluster-gw2-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackuprestoreoncluster-gw2-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:8 Executing query select dictGet('LocalClickHouse_cache_', 'Int64_', toUInt64(2)) on local_node Executing query select dictGet('RedisHash_complex_key_hashed_Int8_', 'Int8_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query SELECT dictGetString('dep_y', 'a', toUInt64(1)) on node1 Executing query select dictGetUInt16('MySQL_cache_', 'UInt16_', toUInt64(2)) on node Executing query select dictGetInt8('RedisHash_complex_key_hashed_Int8_', 'Int8_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGetInt64('LocalClickHouse_cache_', 'Int64_', toUInt64(2)) on local_node Executing query SELECT dictGetString('dep_z', 'a', toUInt64(1)) on node1 Executing query select dictHas('MySQL_cache_', toUInt64(2)) on node Executing query select dictHas('RedisHash_complex_key_hashed_Int8_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictHas('LocalClickHouse_cache_', toUInt64(2)) on local_node Executing query SELECT dictGetString('dep_x', 'a', toUInt64(3)) on node1 Executing query select dictGetOrDefault('MySQL_cache_', 'UInt16_', toUInt64(9999999), toUInt16(66)) on node Executing query select dictGetOrDefault('RedisHash_complex_key_hashed_Int8_', 'Int8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt8(-55)) on node Executing query select dictGetOrDefault('LocalClickHouse_cache_', 'Int64_', toUInt64(9999999), toInt64(-88)) on local_node Executing query SELECT dictGetString('dep_y', 'a', toUInt64(3)) on node1 Executing query select dictGetUInt16OrDefault('MySQL_cache_', 'UInt16_', toUInt64(9999999), toUInt16(66)) on node run container_id:roottestbackuprestoreoncluster-gw2-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackuprestoreoncluster-gw2-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:8 Executing query select dictGetInt8OrDefault('RedisHash_complex_key_hashed_Int8_', 'Int8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt8(-55)) on node Executing query select dictGetInt64OrDefault('LocalClickHouse_cache_', 'Int64_', toUInt64(9999999), toInt64(-88)) on local_node Executing query SELECT dictGetString('dep_z', 'a', toUInt64(3)) on node1 Executing query select dictGet('MySQL_cache_', 'UInt32_', toUInt64(2)) on node Executing query system reload dictionary RedisHash_complex_key_cache_Int8_ on node Executing query select dictGet('LocalClickHouse_cache_', 'UUID_', toUInt64(2)) on local_node Executing query select dictGetUInt32('MySQL_cache_', 'UInt32_', toUInt64(2)) on node Executing query INSERT INTO test.elements VALUES (3, 'fire', 30, 8) on node1 Executing query select dictGet('RedisHash_complex_key_cache_Int8_', 'Int8_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGetUUID('LocalClickHouse_cache_', 'UUID_', toUInt64(2)) on local_node Executing query SELECT dictHas('dep_x', toUInt64(3)) on node1 Executing query select dictHas('MySQL_cache_', toUInt64(2)) on node Executing query select dictGetInt8('RedisHash_complex_key_cache_Int8_', 'Int8_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictHas('LocalClickHouse_cache_', toUInt64(2)) on local_node run container_id:roottestbackuprestoreoncluster-gw2-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackuprestoreoncluster-gw2-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:8 Executing query select dictGetOrDefault('MySQL_cache_', 'UInt32_', toUInt64(9999999), toUInt32(77)) on node Executing query select dictHas('RedisHash_complex_key_cache_Int8_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGetOrDefault('LocalClickHouse_cache_', 'UUID_', toUInt64(9999999), toUUID('550e8400-0000-0000-0000-000000000000')) on local_node Executing query select dictGetUInt32OrDefault('MySQL_cache_', 'UInt32_', toUInt64(9999999), toUInt32(77)) on node Executing query select dictGetOrDefault('RedisHash_complex_key_cache_Int8_', 'Int8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt8(-55)) on node Executing query select dictGetUUIDOrDefault('LocalClickHouse_cache_', 'UUID_', toUInt64(9999999), toUUID('550e8400-0000-0000-0000-000000000000')) on local_node Executing query select dictGet('MySQL_cache_', 'UInt64_', toUInt64(2)) on node Executing query select dictGetInt8OrDefault('RedisHash_complex_key_cache_Int8_', 'Int8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt8(-55)) on node Executing query select dictGet('LocalClickHouse_cache_', 'Date_', toUInt64(2)) on local_node Executing query select dictGetUInt64('MySQL_cache_', 'UInt64_', toUInt64(2)) on node Executing query select dictGet('RedisHash_complex_key_cache_Int8_', 'Int8_', tuple(toUInt64(2),toString('qwerty2'))) on node run container_id:roottestbackuprestoreoncluster-gw2-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackuprestoreoncluster-gw2-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:8 Executing query select dictGetDate('LocalClickHouse_cache_', 'Date_', toUInt64(2)) on local_node Executing query select dictHas('MySQL_cache_', toUInt64(2)) on node Executing query select dictGetInt8('RedisHash_complex_key_cache_Int8_', 'Int8_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictHas('LocalClickHouse_cache_', toUInt64(2)) on local_node Executing query select dictGetOrDefault('MySQL_cache_', 'UInt64_', toUInt64(9999999), toUInt64(88)) on node Executing query select dictHas('RedisHash_complex_key_cache_Int8_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGetOrDefault('LocalClickHouse_cache_', 'Date_', toUInt64(9999999), toDate('2018-12-30')) on local_node Executing query select dictGetUInt64OrDefault('MySQL_cache_', 'UInt64_', toUInt64(9999999), toUInt64(88)) on node Executing query select dictGetOrDefault('RedisHash_complex_key_cache_Int8_', 'Int8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt8(-55)) on node Executing query select dictGetDateOrDefault('LocalClickHouse_cache_', 'Date_', toUInt64(9999999), toDate('2018-12-30')) on local_node Executing query SELECT dictHas('dep_x', toUInt64(3)) on node1 Executing query select dictGet('MySQL_cache_', 'Int8_', toUInt64(2)) on node run container_id:roottestbackuprestoreoncluster-gw2-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackuprestoreoncluster-gw2-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:8 Executing query select dictGetInt8OrDefault('RedisHash_complex_key_cache_Int8_', 'Int8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt8(-55)) on node Executing query select dictGet('LocalClickHouse_cache_', 'DateTime_', toUInt64(2)) on local_node Executing query select dictGetInt8('MySQL_cache_', 'Int8_', toUInt64(2)) on node Executing query system reload dictionary RedisHash_complex_key_direct_Int8_ on node Executing query select dictGetDateTime('LocalClickHouse_cache_', 'DateTime_', toUInt64(2)) on local_node Executing query select dictHas('MySQL_cache_', toUInt64(2)) on node Executing query select dictGet('RedisHash_complex_key_direct_Int8_', 'Int8_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictHas('LocalClickHouse_cache_', toUInt64(2)) on local_node Executing query select dictGetOrDefault('MySQL_cache_', 'Int8_', toUInt64(9999999), toInt8(-55)) on node Executing query select dictGetInt8('RedisHash_complex_key_direct_Int8_', 'Int8_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGetOrDefault('LocalClickHouse_cache_', 'DateTime_', toUInt64(9999999), toDateTime('2018-12-30 00:00:00')) on local_node run container_id:roottestbackuprestoreoncluster-gw2-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackuprestoreoncluster-gw2-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select dictGetInt8OrDefault('MySQL_cache_', 'Int8_', toUInt64(9999999), toInt8(-55)) on node Stdout:8 Executing query select dictHas('RedisHash_complex_key_direct_Int8_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGetDateTimeOrDefault('LocalClickHouse_cache_', 'DateTime_', toUInt64(9999999), toDateTime('2018-12-30 00:00:00')) on local_node Executing query select dictGet('MySQL_cache_', 'Int16_', toUInt64(2)) on node Executing query select dictGetOrDefault('RedisHash_complex_key_direct_Int8_', 'Int8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt8(-55)) on node Executing query select dictGet('LocalClickHouse_cache_', 'String_', toUInt64(2)) on local_node Executing query select dictGetInt16('MySQL_cache_', 'Int16_', toUInt64(2)) on node Executing query select dictGetInt8OrDefault('RedisHash_complex_key_direct_Int8_', 'Int8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt8(-55)) on node Executing query select dictGetString('LocalClickHouse_cache_', 'String_', toUInt64(2)) on local_node Executing query select dictHas('MySQL_cache_', toUInt64(2)) on node Executing query select dictGet('RedisHash_complex_key_direct_Int8_', 'Int8_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictHas('LocalClickHouse_cache_', toUInt64(2)) on local_node run container_id:roottestbackuprestoreoncluster-gw2-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackuprestoreoncluster-gw2-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select dictGetOrDefault('MySQL_cache_', 'Int16_', toUInt64(9999999), toInt16(-66)) on node Stdout:8 Executing query SELECT dictHas('dep_x', toUInt64(3)) on node1 Executing query select dictGetInt8('RedisHash_complex_key_direct_Int8_', 'Int8_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGetOrDefault('LocalClickHouse_cache_', 'String_', toUInt64(9999999), toString('hi')) on local_node Executing query select dictGetInt16OrDefault('MySQL_cache_', 'Int16_', toUInt64(9999999), toInt16(-66)) on node Executing query select dictHas('RedisHash_complex_key_direct_Int8_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGetStringOrDefault('LocalClickHouse_cache_', 'String_', toUInt64(9999999), toString('hi')) on local_node Executing query select dictGet('MySQL_cache_', 'Int32_', toUInt64(2)) on node Executing query select dictGetOrDefault('RedisHash_complex_key_direct_Int8_', 'Int8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt8(-55)) on node Executing query select dictGet('LocalClickHouse_cache_', 'Float32_', toUInt64(2)) on local_node Executing query select dictGetInt32('MySQL_cache_', 'Int32_', toUInt64(2)) on node Executing query select dictGetInt8OrDefault('RedisHash_complex_key_direct_Int8_', 'Int8_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toInt8(-55)) on node run container_id:roottestbackuprestoreoncluster-gw2-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackuprestoreoncluster-gw2-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select dictGetFloat32('LocalClickHouse_cache_', 'Float32_', toUInt64(2)) on local_node Stdout:8 Executing query select dictHas('MySQL_cache_', toUInt64(2)) on node Executing query system reload dictionaries on node Executing query select dictHas('LocalClickHouse_cache_', toUInt64(2)) on local_node Executing query select dictGetOrDefault('MySQL_cache_', 'Int32_', toUInt64(9999999), toInt32(-77)) on node [gw0] PASSED test_dictionaries_redis/test.py::test_redis_dictionaries[Int8_] test_dictionaries_redis/test.py::test_redis_dictionaries[String_] Run test with id: 11 Executing query system reload dictionaries on node Executing query select dictGetOrDefault('LocalClickHouse_cache_', 'Float32_', toUInt64(9999999), toFloat32(555.11)) on local_node Executing query select dictGetInt32OrDefault('MySQL_cache_', 'Int32_', toUInt64(9999999), toInt32(-77)) on node Executing query system reload dictionary RedisSimple_flat_String_ on node Executing query select dictGet('MySQL_cache_', 'Int64_', toUInt64(2)) on node Executing query select dictGetFloat32OrDefault('LocalClickHouse_cache_', 'Float32_', toUInt64(9999999), toFloat32(555.11)) on local_node Executing query select dictGet('RedisSimple_flat_String_', 'String_', toUInt64(1)) on node run container_id:roottestbackuprestoreoncluster-gw2-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackuprestoreoncluster-gw2-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query SELECT status FROM system.backups WHERE id='8d1d19c3-4848-4281-89a6-37dd933e9f7b' AND status == 'CREATING_BACKUP' on node1 Executing query select dictGetInt64('MySQL_cache_', 'Int64_', toUInt64(2)) on node Executing query select dictGet('LocalClickHouse_cache_', 'Float64_', toUInt64(2)) on local_node Executing query select dictGetString('RedisSimple_flat_String_', 'String_', toUInt64(1)) on node Executing query SELECT status FROM system.backups WHERE id='8d1d19c3-4848-4281-89a6-37dd933e9f7b' on node1 Executing query SELECT dictHas('dep_x', toUInt64(3)) on node1 Executing query select dictHas('MySQL_cache_', toUInt64(2)) on node Executing query select dictGetFloat64('LocalClickHouse_cache_', 'Float64_', toUInt64(2)) on local_node Executing query select dictHas('RedisSimple_flat_String_', toUInt64(1)) on node run container_id:roottestbackuprestoreoncluster-gw2-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackuprestoreoncluster-gw2-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select dictGetOrDefault('MySQL_cache_', 'Int64_', toUInt64(9999999), toInt64(-88)) on node No clickhouse process running. Start new one. http://localhost:None "POST /v1.46/containers/roottestbackuprestoreoncluster-gw2-node2-1/exec HTTP/1.1" 201 74 http://localhost:None "POST /v1.46/exec/9793705c9705a02f430e61598ffe799a8089368de95a4bf812a53554cdd7f8f8/start HTTP/1.1" 200 0 http://localhost:None "GET /v1.46/exec/9793705c9705a02f430e61598ffe799a8089368de95a4bf812a53554cdd7f8f8/json HTTP/1.1" 200 586 Executing query select dictHas('LocalClickHouse_cache_', toUInt64(2)) on local_node Executing query select dictGetOrDefault('RedisSimple_flat_String_', 'String_', toUInt64(9999999), toString('hi')) on node Executing query select dictGetInt64OrDefault('MySQL_cache_', 'Int64_', toUInt64(9999999), toInt64(-88)) on node Executing query select dictGetOrDefault('LocalClickHouse_cache_', 'Float64_', toUInt64(9999999), toFloat64(777.11)) on local_node Executing query select dictGetStringOrDefault('RedisSimple_flat_String_', 'String_', toUInt64(9999999), toString('hi')) on node Executing query select dictGet('MySQL_cache_', 'UUID_', toUInt64(2)) on node Executing query select dictGetFloat64OrDefault('LocalClickHouse_cache_', 'Float64_', toUInt64(9999999), toFloat64(777.11)) on local_node Executing query select dictGet('RedisSimple_flat_String_', 'String_', toUInt64(2)) on node Executing query select dictGetUUID('MySQL_cache_', 'UUID_', toUInt64(2)) on node Executing query select dictGet('LocalClickHouse_cache_', 'ParentKeyField', toUInt64(2)) on local_node Executing query select dictGetString('RedisSimple_flat_String_', 'String_', toUInt64(2)) on node run container_id:roottestbackuprestoreoncluster-gw2-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackuprestoreoncluster-gw2-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:877 Clickhouse process running. run container_id:roottestbackuprestoreoncluster-gw2-node2-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestbackuprestoreoncluster-gw2-node2-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:877 Executing query select 20 on node2 Executing query select dictHas('MySQL_cache_', toUInt64(2)) on node Executing query select dictHas('RedisSimple_flat_String_', toUInt64(2)) on node Executing query select dictGetUInt64('LocalClickHouse_cache_', 'ParentKeyField', toUInt64(2)) on local_node Executing query select dictGetOrDefault('MySQL_cache_', 'UUID_', toUInt64(9999999), toUUID('550e8400-0000-0000-0000-000000000000')) on node Executing query select dictGetOrDefault('RedisSimple_flat_String_', 'String_', toUInt64(9999999), toString('hi')) on node Executing query select dictHas('LocalClickHouse_cache_', toUInt64(2)) on local_node Executing query select dictGetUUIDOrDefault('MySQL_cache_', 'UUID_', toUInt64(9999999), toUUID('550e8400-0000-0000-0000-000000000000')) on node Executing query select dictGetOrDefault('LocalClickHouse_cache_', 'ParentKeyField', toUInt64(9999999), toUInt64(444)) on local_node Executing query select dictGetStringOrDefault('RedisSimple_flat_String_', 'String_', toUInt64(9999999), toString('hi')) on node Executing query select 20 on node2 Executing query SELECT dictHas('dep_x', toUInt64(3)) on node1 Executing query select dictGet('MySQL_cache_', 'Date_', toUInt64(2)) on node Executing query select dictGetUInt64OrDefault('LocalClickHouse_cache_', 'ParentKeyField', toUInt64(9999999), toUInt64(444)) on local_node Executing query system reload dictionary RedisSimple_hashed_String_ on node Executing query select dictGetHierarchy('LocalClickHouse_cache_', toUInt64(1)) on local_node Executing query select dictGet('RedisSimple_hashed_String_', 'String_', toUInt64(1)) on node Executing query select dictGetDate('MySQL_cache_', 'Date_', toUInt64(2)) on node Executing query select 20 on node2 Executing query select dictGetString('RedisSimple_hashed_String_', 'String_', toUInt64(1)) on node Executing query select dictHas('MySQL_cache_', toUInt64(2)) on node Executing query select dictGetHierarchy('LocalClickHouse_cache_', toUInt64(2)) on local_node Executing query select dictHas('RedisSimple_hashed_String_', toUInt64(1)) on node Executing query select dictIsIn('LocalClickHouse_cache_', toUInt64(1), toUInt64(2)) on local_node Executing query select dictGetOrDefault('MySQL_cache_', 'Date_', toUInt64(9999999), toDate('2018-12-30')) on node Executing query select 20 on node2 Executing query select dictGetOrDefault('RedisSimple_hashed_String_', 'String_', toUInt64(9999999), toString('hi')) on node Executing query select dictIsIn('LocalClickHouse_cache_', toUInt64(2), toUInt64(1)) on local_node Executing query select dictGetDateOrDefault('MySQL_cache_', 'Date_', toUInt64(9999999), toDate('2018-12-30')) on node Executing query select dictGetStringOrDefault('RedisSimple_hashed_String_', 'String_', toUInt64(9999999), toString('hi')) on node [gw5] PASSED test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_simple[cache] test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_simple[direct] Executing query system reload dictionaries on local_node Executing query select dictGet('MySQL_cache_', 'DateTime_', toUInt64(2)) on node Executing query select dictGet('RedisSimple_hashed_String_', 'String_', toUInt64(2)) on node Executing query select 20 on node2 Executing query select dictGet('LocalClickHouse_direct_', 'UInt8_', toUInt64(1)) on local_node Executing query select dictGetDateTime('MySQL_cache_', 'DateTime_', toUInt64(2)) on node Executing query select dictGetString('RedisSimple_hashed_String_', 'String_', toUInt64(2)) on node Executing query select dictGetUInt8('LocalClickHouse_direct_', 'UInt8_', toUInt64(1)) on local_node Executing query select dictHas('MySQL_cache_', toUInt64(2)) on node Executing query SELECT dictHas('dep_x', toUInt64(3)) on node1 Executing query select dictHas('RedisSimple_hashed_String_', toUInt64(2)) on node Executing query select dictGetOrDefault('MySQL_cache_', 'DateTime_', toUInt64(9999999), toDateTime('2018-12-30 00:00:00')) on node Executing query select dictHas('LocalClickHouse_direct_', toUInt64(1)) on local_node Executing query select 20 on node2 Executing query select dictGetOrDefault('RedisSimple_hashed_String_', 'String_', toUInt64(9999999), toString('hi')) on node Executing query select dictGetDateTimeOrDefault('MySQL_cache_', 'DateTime_', toUInt64(9999999), toDateTime('2018-12-30 00:00:00')) on node Executing query DROP TABLE tbl ON CLUSTER 'cluster' SYNC on node1 Executing query select dictGetOrDefault('LocalClickHouse_direct_', 'UInt8_', toUInt64(9999999), toUInt8(55)) on local_node Executing query select dictGetStringOrDefault('RedisSimple_hashed_String_', 'String_', toUInt64(9999999), toString('hi')) on node Executing query select dictGetUInt8OrDefault('LocalClickHouse_direct_', 'UInt8_', toUInt64(9999999), toUInt8(55)) on local_node Executing query select dictGet('MySQL_cache_', 'String_', toUInt64(2)) on node Executing query system reload dictionary RedisSimple_cache_String_ on node Executing query RESTORE TABLE tbl ON CLUSTER 'cluster' FROM Disk('backups', '26') on node1 Executing query select dictGetString('MySQL_cache_', 'String_', toUInt64(2)) on node Executing query select dictGet('LocalClickHouse_direct_', 'UInt16_', toUInt64(1)) on local_node Executing query select dictGet('RedisSimple_cache_String_', 'String_', toUInt64(1)) on node Executing query select dictHas('MySQL_cache_', toUInt64(2)) on node Executing query select dictGetUInt16('LocalClickHouse_direct_', 'UInt16_', toUInt64(1)) on local_node Executing query select dictGetString('RedisSimple_cache_String_', 'String_', toUInt64(1)) on node Executing query SYSTEM SYNC REPLICA tbl on node1 Executing query select dictGetOrDefault('MySQL_cache_', 'String_', toUInt64(9999999), toString('hi')) on node Executing query select dictHas('RedisSimple_cache_String_', toUInt64(1)) on node Executing query select dictHas('LocalClickHouse_direct_', toUInt64(1)) on local_node Executing query SELECT * FROM tbl ORDER BY x on node1 Executing query select dictGetStringOrDefault('MySQL_cache_', 'String_', toUInt64(9999999), toString('hi')) on node Executing query select dictGetOrDefault('RedisSimple_cache_String_', 'String_', toUInt64(9999999), toString('hi')) on node Executing query select dictGetOrDefault('LocalClickHouse_direct_', 'UInt16_', toUInt64(9999999), toUInt16(66)) on local_node Executing query DROP TABLE IF EXISTS tbl ON CLUSTER 'cluster3' SYNC on node1 [gw2] PASSED test_backup_restore_on_cluster/test.py::test_shutdown_waits_for_backup Executing query select dictGet('MySQL_cache_', 'Float32_', toUInt64(2)) on node Executing query SELECT dictHas('dep_x', toUInt64(3)) on node1 Executing query select dictGetStringOrDefault('RedisSimple_cache_String_', 'String_', toUInt64(9999999), toString('hi')) on node Executing query select dictGetUInt16OrDefault('LocalClickHouse_direct_', 'UInt16_', toUInt64(9999999), toUInt16(66)) on local_node Executing query select dictGetFloat32('MySQL_cache_', 'Float32_', toUInt64(2)) on node Executing query SELECT dictGetString('dep_x', 'a', toUInt64(3)) on node1 Executing query DROP TABLE IF EXISTS tbl2 ON CLUSTER 'cluster3' SYNC on node1 Executing query select dictGet('RedisSimple_cache_String_', 'String_', toUInt64(2)) on node Executing query select dictGet('LocalClickHouse_direct_', 'UInt32_', toUInt64(1)) on local_node Executing query SELECT dictGetString('dep_y', 'a', toUInt64(3)) on node1 Executing query select dictHas('MySQL_cache_', toUInt64(2)) on node Executing query select dictGetString('RedisSimple_cache_String_', 'String_', toUInt64(2)) on node Executing query select dictGetUInt32('LocalClickHouse_direct_', 'UInt32_', toUInt64(1)) on local_node Executing query DROP DATABASE IF EXISTS mydb ON CLUSTER 'cluster3' SYNC on node1 Executing query SELECT dictGetString('dep_z', 'a', toUInt64(3)) on node1 Executing query select dictGetOrDefault('MySQL_cache_', 'Float32_', toUInt64(9999999), toFloat32(555.11)) on node Executing query select dictHas('RedisSimple_cache_String_', toUInt64(2)) on node Executing query select dictHas('LocalClickHouse_direct_', toUInt64(1)) on local_node Executing query INSERT INTO test.elements VALUES (4, 'ether', 404, 0.001) on node1 Executing query DROP DATABASE IF EXISTS mydb2 ON CLUSTER 'cluster3' SYNC on node1 Executing query select dictGetFloat32OrDefault('MySQL_cache_', 'Float32_', toUInt64(9999999), toFloat32(555.11)) on node Executing query select dictGetOrDefault('RedisSimple_cache_String_', 'String_', toUInt64(9999999), toString('hi')) on node Executing query SELECT dictHas('dep_y', toUInt64(4)) on node1 Executing query select dictGetOrDefault('LocalClickHouse_direct_', 'UInt32_', toUInt64(9999999), toUInt32(77)) on local_node Executing query DROP USER IF EXISTS u1, u2 ON CLUSTER 'cluster3' on node1 Executing query select dictGet('MySQL_cache_', 'Float64_', toUInt64(2)) on node Executing query select dictGetStringOrDefault('RedisSimple_cache_String_', 'String_', toUInt64(9999999), toString('hi')) on node Executing query select dictGetUInt32OrDefault('LocalClickHouse_direct_', 'UInt32_', toUInt64(9999999), toUInt32(77)) on local_node Executing query system reload dictionary RedisSimple_direct_String_ on node Executing query select dictGetFloat64('MySQL_cache_', 'Float64_', toUInt64(2)) on node test_backup_restore_on_cluster/test.py::test_system_functions Executing query CREATE FUNCTION linear_equation AS (x, k, b) -> k*x + b; on node1 Executing query select dictGet('LocalClickHouse_direct_', 'UInt64_', toUInt64(1)) on local_node Executing query select dictGet('RedisSimple_direct_String_', 'String_', toUInt64(1)) on node Executing query CREATE FUNCTION parity_str AS (n) -> if(n % 2, 'odd', 'even'); on node1 Executing query select dictHas('MySQL_cache_', toUInt64(2)) on node Executing query select dictGetString('RedisSimple_direct_String_', 'String_', toUInt64(1)) on node Executing query BACKUP TABLE system.functions ON CLUSTER 'cluster' TO Disk('backups', '27') on node1 Executing query select dictGetUInt64('LocalClickHouse_direct_', 'UInt64_', toUInt64(1)) on local_node Executing query select dictGetOrDefault('MySQL_cache_', 'Float64_', toUInt64(9999999), toFloat64(777.11)) on node Executing query select dictHas('RedisSimple_direct_String_', toUInt64(1)) on node Executing query select dictHas('LocalClickHouse_direct_', toUInt64(1)) on local_node Executing query select dictGetFloat64OrDefault('MySQL_cache_', 'Float64_', toUInt64(9999999), toFloat64(777.11)) on node Executing query DROP FUNCTION linear_equation on node1 Executing query select dictGetOrDefault('RedisSimple_direct_String_', 'String_', toUInt64(9999999), toString('hi')) on node Executing query select dictGetOrDefault('LocalClickHouse_direct_', 'UInt64_', toUInt64(9999999), toUInt64(88)) on local_node Executing query select dictGet('MySQL_cache_', 'ParentKeyField', toUInt64(2)) on node Executing query DROP FUNCTION parity_str on node1 Executing query select dictGetStringOrDefault('RedisSimple_direct_String_', 'String_', toUInt64(9999999), toString('hi')) on node Executing query select dictGetUInt64OrDefault('LocalClickHouse_direct_', 'UInt64_', toUInt64(9999999), toUInt64(88)) on local_node Executing query select dictGetUInt64('MySQL_cache_', 'ParentKeyField', toUInt64(2)) on node Executing query SELECT name FROM system.functions WHERE name='parity_str' on node2 Executing query select dictGet('RedisSimple_direct_String_', 'String_', toUInt64(2)) on node Executing query select dictGet('LocalClickHouse_direct_', 'Int8_', toUInt64(1)) on local_node Executing query select dictHas('MySQL_cache_', toUInt64(2)) on node Executing query SELECT dictHas('dep_y', toUInt64(4)) on node1 Executing query RESTORE TABLE system.functions ON CLUSTER 'cluster' FROM Disk('backups', '27') on node1 Executing query select dictGetString('RedisSimple_direct_String_', 'String_', toUInt64(2)) on node Executing query select dictGetInt8('LocalClickHouse_direct_', 'Int8_', toUInt64(1)) on local_node Executing query select dictGetOrDefault('MySQL_cache_', 'ParentKeyField', toUInt64(9999999), toUInt64(444)) on node Executing query select dictHas('RedisSimple_direct_String_', toUInt64(2)) on node Executing query SELECT number, linear_equation(number, 2, 1) FROM numbers(3) on node1 Executing query select dictHas('LocalClickHouse_direct_', toUInt64(1)) on local_node Executing query select dictGetUInt64OrDefault('MySQL_cache_', 'ParentKeyField', toUInt64(9999999), toUInt64(444)) on node Executing query select dictGetOrDefault('RedisSimple_direct_String_', 'String_', toUInt64(9999999), toString('hi')) on node Executing query SELECT number, parity_str(number) FROM numbers(3) on node1 Executing query select dictGetOrDefault('LocalClickHouse_direct_', 'Int8_', toUInt64(9999999), toInt8(-55)) on local_node Executing query select dictGetHierarchy('MySQL_cache_', toUInt64(1)) on node Executing query select dictGetStringOrDefault('RedisSimple_direct_String_', 'String_', toUInt64(9999999), toString('hi')) on node Executing query SELECT number, linear_equation(number, 2, 1) FROM numbers(3) on node2 Executing query select dictGetInt8OrDefault('LocalClickHouse_direct_', 'Int8_', toUInt64(9999999), toInt8(-55)) on local_node Executing query select dictGetHierarchy('MySQL_cache_', toUInt64(2)) on node Executing query system reload dictionary RedisHash_complex_key_hashed_String_ on node Executing query SELECT number, parity_str(number) FROM numbers(3) on node2 Executing query select dictIsIn('MySQL_cache_', toUInt64(1), toUInt64(2)) on node Executing query select dictGet('LocalClickHouse_direct_', 'Int16_', toUInt64(1)) on local_node Executing query select dictGet('RedisHash_complex_key_hashed_String_', 'String_', tuple(toUInt64(1),toString('world'))) on node Executing query SELECT name FROM system.functions WHERE name='parity_str' on node2 Executing query select dictIsIn('MySQL_cache_', toUInt64(2), toUInt64(1)) on node Executing query select dictGetInt16('LocalClickHouse_direct_', 'Int16_', toUInt64(1)) on local_node Executing query select dictGetString('RedisHash_complex_key_hashed_String_', 'String_', tuple(toUInt64(1),toString('world'))) on node Executing query SELECT number, parity_str(number) FROM numbers(3) on node2 [gw3] PASSED test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_simple[cache] test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_simple[direct] Executing query system reload dictionaries on node Executing query select dictHas('RedisHash_complex_key_hashed_String_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictHas('LocalClickHouse_direct_', toUInt64(1)) on local_node Executing query SELECT dictHas('dep_y', toUInt64(4)) on node1 Executing query DROP FUNCTION linear_equation on node1 Executing query select dictGet('MySQL_direct_', 'UInt8_', toUInt64(1)) on node Executing query select dictGetOrDefault('RedisHash_complex_key_hashed_String_', 'String_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toString('hi')) on node Executing query SELECT dictGetString('dep_x', 'a', toUInt64(4)) on node1 Executing query select dictGetOrDefault('LocalClickHouse_direct_', 'Int16_', toUInt64(9999999), toInt16(-66)) on local_node Executing query DROP FUNCTION parity_str on node1 Executing query select dictGetUInt8('MySQL_direct_', 'UInt8_', toUInt64(1)) on node Executing query select dictGetStringOrDefault('RedisHash_complex_key_hashed_String_', 'String_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toString('hi')) on node Executing query SELECT dictGetString('dep_y', 'a', toUInt64(4)) on node1 Executing query DROP TABLE IF EXISTS tbl ON CLUSTER 'cluster3' SYNC on node1 [gw2] PASSED test_backup_restore_on_cluster/test.py::test_system_functions Executing query select dictGetInt16OrDefault('LocalClickHouse_direct_', 'Int16_', toUInt64(9999999), toInt16(-66)) on local_node Executing query select dictHas('MySQL_direct_', toUInt64(1)) on node Executing query select dictGet('RedisHash_complex_key_hashed_String_', 'String_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query SELECT dictGetString('dep_z', 'a', toUInt64(4)) on node1 Executing query select dictGet('LocalClickHouse_direct_', 'Int32_', toUInt64(1)) on local_node Executing query DROP TABLE IF EXISTS tbl2 ON CLUSTER 'cluster3' SYNC on node1 Connection dropped: socket connection error: None Executing query select dictGetOrDefault('MySQL_direct_', 'UInt8_', toUInt64(9999999), toUInt8(55)) on node Executing query select dictGetString('RedisHash_complex_key_hashed_String_', 'String_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query DROP TABLE IF EXISTS test.elements; on node1 Executing query select dictGetInt32('LocalClickHouse_direct_', 'Int32_', toUInt64(1)) on local_node run container_id:roottestasyncloaddatabases-gw7-node1-1 detach:False nothrow:True cmd: ['bash', '-c', 'ps -C clickhouse'] Command:[docker exec -u root roottestasyncloaddatabases-gw7-node1-1 bash -c ps -C clickhouse] Executing query DROP DATABASE IF EXISTS mydb ON CLUSTER 'cluster3' SYNC on node1 Stdout: PID TTY TIME CMD Stdout: 774 ? 00:00:17 clickhouse run container_id:roottestasyncloaddatabases-gw7-node1-1 detach:False nothrow:False cmd: ['bash', '-c', 'pkill clickhouse'] Command:[docker exec -u root roottestasyncloaddatabases-gw7-node1-1 bash -c pkill clickhouse] Executing query select dictGetUInt8OrDefault('MySQL_direct_', 'UInt8_', toUInt64(9999999), toUInt8(55)) on node run container_id:roottestasyncloaddatabases-gw7-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw7-node1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select dictHas('RedisHash_complex_key_hashed_String_', tuple(toUInt64(2),toString('qwerty2'))) on node Stdout:774 Executing query select dictHas('LocalClickHouse_direct_', toUInt64(1)) on local_node Executing query select dictGet('MySQL_direct_', 'UInt16_', toUInt64(1)) on node Executing query DROP DATABASE IF EXISTS mydb2 ON CLUSTER 'cluster3' SYNC on node1 Executing query select dictGetOrDefault('RedisHash_complex_key_hashed_String_', 'String_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toString('hi')) on node Executing query select dictGetUInt16('MySQL_direct_', 'UInt16_', toUInt64(1)) on node Executing query select dictGetOrDefault('LocalClickHouse_direct_', 'Int32_', toUInt64(9999999), toInt32(-77)) on local_node Executing query select dictGetStringOrDefault('RedisHash_complex_key_hashed_String_', 'String_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toString('hi')) on node Executing query DROP USER IF EXISTS u1, u2 ON CLUSTER 'cluster3' on node1 Executing query select dictHas('MySQL_direct_', toUInt64(1)) on node Executing query select dictGetInt32OrDefault('LocalClickHouse_direct_', 'Int32_', toUInt64(9999999), toInt32(-77)) on local_node Executing query system reload dictionary RedisHash_complex_key_cache_String_ on node test_backup_restore_on_cluster/test.py::test_system_users Executing query CREATE USER u1 SETTINGS custom_a=123 on node1 run container_id:roottestasyncloaddatabases-gw7-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw7-node1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:774 Executing query select dictGet('RedisHash_complex_key_cache_String_', 'String_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGetOrDefault('MySQL_direct_', 'UInt16_', toUInt64(9999999), toUInt16(66)) on node Executing query select dictGet('LocalClickHouse_direct_', 'Int64_', toUInt64(1)) on local_node Executing query GRANT SELECT ON tbl TO u1 on node1 Executing query select dictGetUInt16OrDefault('MySQL_direct_', 'UInt16_', toUInt64(9999999), toUInt16(66)) on node Executing query select dictGetString('RedisHash_complex_key_cache_String_', 'String_', tuple(toUInt64(1),toString('world'))) on node Executing query CREATE USER u2 SETTINGS allow_backup=false on node1 Executing query select dictGetInt64('LocalClickHouse_direct_', 'Int64_', toUInt64(1)) on local_node Executing query GRANT CLUSTER ON *.* TO u2 on node1 Executing query select dictGet('MySQL_direct_', 'UInt32_', toUInt64(1)) on node Executing query select dictHas('RedisHash_complex_key_cache_String_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictHas('LocalClickHouse_direct_', toUInt64(1)) on local_node Executing query BACKUP TABLE system.users ON CLUSTER 'cluster' TO Disk('backups', '28') on node1 Executing query select dictGetUInt32('MySQL_direct_', 'UInt32_', toUInt64(1)) on node Executing query select dictGetOrDefault('RedisHash_complex_key_cache_String_', 'String_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toString('hi')) on node run container_id:roottestasyncloaddatabases-gw7-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw7-node1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select dictGetOrDefault('LocalClickHouse_direct_', 'Int64_', toUInt64(9999999), toInt64(-88)) on local_node Stdout:774 Executing query select dictHas('MySQL_direct_', toUInt64(1)) on node Executing query select dictGetStringOrDefault('RedisHash_complex_key_cache_String_', 'String_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toString('hi')) on node Executing query GRANT BACKUP ON system.users TO u2 on node1 Executing query select dictGetInt64OrDefault('LocalClickHouse_direct_', 'Int64_', toUInt64(9999999), toInt64(-88)) on local_node Executing query select dictGetOrDefault('MySQL_direct_', 'UInt32_', toUInt64(9999999), toUInt32(77)) on node Executing query BACKUP TABLE system.users ON CLUSTER 'cluster' TO Disk('backups', '28') on node1 Executing query select dictGet('RedisHash_complex_key_cache_String_', 'String_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGetUInt32OrDefault('MySQL_direct_', 'UInt32_', toUInt64(9999999), toUInt32(77)) on node Executing query select dictGet('LocalClickHouse_direct_', 'UUID_', toUInt64(1)) on local_node Executing query select dictGetString('RedisHash_complex_key_cache_String_', 'String_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGet('MySQL_direct_', 'UInt64_', toUInt64(1)) on node Executing query DROP USER u1 on node1 Executing query select dictGetUUID('LocalClickHouse_direct_', 'UUID_', toUInt64(1)) on local_node Executing query select dictHas('RedisHash_complex_key_cache_String_', tuple(toUInt64(2),toString('qwerty2'))) on node Connection dropped: socket connection error: None run container_id:roottestasyncloaddatabases-gw7-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw7-node1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:774 Executing query RESTORE TABLE system.users ON CLUSTER 'cluster' FROM Disk('backups', '28') on node1 Executing query select dictGetUInt64('MySQL_direct_', 'UInt64_', toUInt64(1)) on node Executing query select dictGetOrDefault('RedisHash_complex_key_cache_String_', 'String_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toString('hi')) on node Executing query select dictHas('LocalClickHouse_direct_', toUInt64(1)) on local_node Executing query select dictHas('MySQL_direct_', toUInt64(1)) on node Executing query select dictGetStringOrDefault('RedisHash_complex_key_cache_String_', 'String_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toString('hi')) on node Executing query GRANT CREATE USER ON *.* TO u2 on node1 Executing query select dictGetOrDefault('LocalClickHouse_direct_', 'UUID_', toUInt64(9999999), toUUID('550e8400-0000-0000-0000-000000000000')) on local_node Executing query select dictGetOrDefault('MySQL_direct_', 'UInt64_', toUInt64(9999999), toUInt64(88)) on node Executing query RESTORE TABLE system.users ON CLUSTER 'cluster' FROM Disk('backups', '28') on node1 Executing query system reload dictionary RedisHash_complex_key_direct_String_ on node Executing query select dictGetUUIDOrDefault('LocalClickHouse_direct_', 'UUID_', toUInt64(9999999), toUUID('550e8400-0000-0000-0000-000000000000')) on local_node Executing query select dictGet('RedisHash_complex_key_direct_String_', 'String_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGetUInt64OrDefault('MySQL_direct_', 'UInt64_', toUInt64(9999999), toUInt64(88)) on node run container_id:roottestasyncloaddatabases-gw7-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw7-node1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query GRANT SELECT ON tbl TO u2 WITH GRANT OPTION on node1 Stdout:774 Executing query select dictGet('LocalClickHouse_direct_', 'Date_', toUInt64(1)) on local_node Executing query select dictGetString('RedisHash_complex_key_direct_String_', 'String_', tuple(toUInt64(1),toString('world'))) on node Executing query RESTORE TABLE system.users ON CLUSTER 'cluster' FROM Disk('backups', '28') on node1 Executing query select dictGet('MySQL_direct_', 'Int8_', toUInt64(1)) on node Executing query select dictGetDate('LocalClickHouse_direct_', 'Date_', toUInt64(1)) on local_node Executing query select dictHas('RedisHash_complex_key_direct_String_', tuple(toUInt64(1),toString('world'))) on node Executing query select dictGetInt8('MySQL_direct_', 'Int8_', toUInt64(1)) on node Executing query select dictHas('LocalClickHouse_direct_', toUInt64(1)) on local_node Executing query SHOW CREATE USER u1 on node1 Executing query select dictGetOrDefault('RedisHash_complex_key_direct_String_', 'String_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toString('hi')) on node Executing query select dictHas('MySQL_direct_', toUInt64(1)) on node Executing query select dictGetOrDefault('LocalClickHouse_direct_', 'Date_', toUInt64(9999999), toDate('2018-12-30')) on local_node run container_id:roottestasyncloaddatabases-gw7-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw7-node1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query SHOW GRANTS FOR u1 on node1 Executing query select dictGetStringOrDefault('RedisHash_complex_key_direct_String_', 'String_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toString('hi')) on node run container_id:roottestasyncloaddatabases-gw7-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw7-node1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select dictGetOrDefault('MySQL_direct_', 'Int8_', toUInt64(9999999), toInt8(-55)) on node No clickhouse process running. Start new one. http://localhost:None "POST /v1.46/containers/roottestasyncloaddatabases-gw7-node1-1/exec HTTP/1.1" 201 74 http://localhost:None "POST /v1.46/exec/f4f5a2f269d115333c133d4aad734d6b13d2c23db983b3e0bd6f07a0e218ef3c/start HTTP/1.1" 200 0 http://localhost:None "GET /v1.46/exec/f4f5a2f269d115333c133d4aad734d6b13d2c23db983b3e0bd6f07a0e218ef3c/json HTTP/1.1" 200 586 Executing query DROP TABLE IF EXISTS tbl ON CLUSTER 'cluster3' SYNC on node1 [gw2] PASSED test_backup_restore_on_cluster/test.py::test_system_users Executing query select dictGet('RedisHash_complex_key_direct_String_', 'String_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGetDateOrDefault('LocalClickHouse_direct_', 'Date_', toUInt64(9999999), toDate('2018-12-30')) on local_node Executing query select dictGetInt8OrDefault('MySQL_direct_', 'Int8_', toUInt64(9999999), toInt8(-55)) on node Executing query select dictGetString('RedisHash_complex_key_direct_String_', 'String_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGet('LocalClickHouse_direct_', 'DateTime_', toUInt64(1)) on local_node Executing query select dictGet('MySQL_direct_', 'Int16_', toUInt64(1)) on node Executing query DROP TABLE IF EXISTS tbl2 ON CLUSTER 'cluster3' SYNC on node1 Executing query select dictGetInt16('MySQL_direct_', 'Int16_', toUInt64(1)) on node Executing query select dictHas('RedisHash_complex_key_direct_String_', tuple(toUInt64(2),toString('qwerty2'))) on node Executing query select dictGetDateTime('LocalClickHouse_direct_', 'DateTime_', toUInt64(1)) on local_node Executing query DROP DATABASE IF EXISTS mydb ON CLUSTER 'cluster3' SYNC on node1 run container_id:roottestasyncloaddatabases-gw7-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw7-node1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select dictHas('MySQL_direct_', toUInt64(1)) on node Stdout:1581 Clickhouse process running. run container_id:roottestasyncloaddatabases-gw7-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw7-node1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select dictGetOrDefault('RedisHash_complex_key_direct_String_', 'String_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toString('hi')) on node Stdout:1581 Executing query select 20 on node1 Executing query select dictHas('LocalClickHouse_direct_', toUInt64(1)) on local_node Executing query DROP DATABASE IF EXISTS mydb2 ON CLUSTER 'cluster3' SYNC on node1 Executing query select dictGetOrDefault('MySQL_direct_', 'Int16_', toUInt64(9999999), toInt16(-66)) on node Executing query select dictGetStringOrDefault('RedisHash_complex_key_direct_String_', 'String_', tuple(toUInt64(9999999),toString('xxxxxxxxx')), toString('hi')) on node Executing query select dictGetOrDefault('LocalClickHouse_direct_', 'DateTime_', toUInt64(9999999), toDateTime('2018-12-30 00:00:00')) on local_node Executing query DROP USER IF EXISTS u1, u2 ON CLUSTER 'cluster3' on node1 Executing query select dictGetInt16OrDefault('MySQL_direct_', 'Int16_', toUInt64(9999999), toInt16(-66)) on node Executing query system reload dictionaries on node Executing query select dictGetDateTimeOrDefault('LocalClickHouse_direct_', 'DateTime_', toUInt64(9999999), toDateTime('2018-12-30 00:00:00')) on local_node Executing query select 20 on node1 test_backup_restore_on_cluster/test.py::test_table_in_replicated_database_with_not_synced_def Executing query CREATE DATABASE mydb ON CLUSTER 'cluster' ENGINE=Replicated('/clickhouse/path/','{shard}','{replica}') on node1 Executing query select dictGet('MySQL_direct_', 'Int32_', toUInt64(1)) on node Command:[docker compose --env-file /ClickHouse/tests/integration/test_dictionaries_redis/_instances-0-gw0/.env --project-name roottestdictionariesredis-gw0 --file /ClickHouse/tests/integration/test_dictionaries_redis/_instances-0-gw0/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_redis.yml stop --timeout 20] [gw0] PASSED test_dictionaries_redis/test.py::test_redis_dictionaries[String_] Executing query select dictGet('LocalClickHouse_direct_', 'String_', toUInt64(1)) on local_node Executing query select dictGetInt32('MySQL_direct_', 'Int32_', toUInt64(1)) on node Executing query CREATE TABLE mydb.tbl (x UInt8, y String) ENGINE=ReplicatedMergeTree ORDER BY tuple() on node1 Executing query select dictGetString('LocalClickHouse_direct_', 'String_', toUInt64(1)) on local_node Executing query select 20 on node1 Executing query select dictHas('MySQL_direct_', toUInt64(1)) on node Executing query select dictHas('LocalClickHouse_direct_', toUInt64(1)) on local_node Executing query ALTER TABLE mydb.tbl MODIFY COLUMN x String on node1 Stderr: node1 Skipped - Image is already being pulled by node2 Stderr: zoo1 Skipped - Image is already being pulled by node2 Stderr: zoo2 Skipped - Image is already being pulled by node2 Stderr: zoo3 Skipped - Image is already being pulled by node2 Stderr: node2 Pulling Stderr: azurite1 Pulling Stderr: node2 Pulled Stderr: f18232174bc9 Pulling fs layer Stderr: cb2bde55f71f Pulling fs layer Stderr: 9d0e0719fbe0 Pulling fs layer Stderr: 6f063dbd7a5d Pulling fs layer Stderr: f9e3e3d8f042 Pulling fs layer Stderr: a39ef2f62dc8 Pulling fs layer Stderr: 9a21c6b23f0e Pulling fs layer Stderr: efeb7b313b67 Pulling fs layer Stderr: 6fef65209747 Pulling fs layer Stderr: 3d377e512a83 Pulling fs layer Stderr: 6f063dbd7a5d Waiting Stderr: f9e3e3d8f042 Waiting Stderr: a39ef2f62dc8 Waiting Stderr: 9a21c6b23f0e Waiting Stderr: efeb7b313b67 Waiting Stderr: 6fef65209747 Waiting Stderr: 3d377e512a83 Waiting Stderr: 9d0e0719fbe0 Downloading [> ] 15.58kB/1.261MB Stderr: f18232174bc9 Downloading [> ] 48.34kB/3.642MB Stderr: 9d0e0719fbe0 Downloading [==============> ] 358.7kB/1.261MB Stderr: cb2bde55f71f Downloading [> ] 506.1kB/50.34MB Stderr: f18232174bc9 Downloading [=======> ] 538.9kB/3.642MB Stderr: 9d0e0719fbe0 Verifying Checksum Stderr: 9d0e0719fbe0 Download complete Executing query select dictGetOrDefault('MySQL_direct_', 'Int32_', toUInt64(9999999), toInt32(-77)) on node Stderr: cb2bde55f71f Downloading [==> ] 2.03MB/50.34MB Stderr: f18232174bc9 Downloading [====================================> ] 2.652MB/3.642MB Stderr: f18232174bc9 Verifying Checksum Stderr: f18232174bc9 Download complete Stderr: f18232174bc9 Extracting [> ] 65.54kB/3.642MB Stderr: cb2bde55f71f Downloading [=====> ] 5.077MB/50.34MB Stderr: f18232174bc9 Extracting [==================================================>] 3.642MB/3.642MB Stderr: f18232174bc9 Extracting [==================================================>] 3.642MB/3.642MB Stderr: f18232174bc9 Pull complete Stderr: cb2bde55f71f Downloading [============> ] 12.19MB/50.34MB Stderr: cb2bde55f71f Downloading [=======================> ] 23.36MB/50.34MB Stderr: cb2bde55f71f Downloading [===================================> ] 36.06MB/50.34MB Stderr: cb2bde55f71f Downloading [===============================================> ] 47.74MB/50.34MB Stderr: cb2bde55f71f Verifying Checksum Stderr: cb2bde55f71f Download complete Stderr: cb2bde55f71f Extracting [> ] 524.3kB/50.34MB Stderr: cb2bde55f71f Extracting [=======> ] 7.864MB/50.34MB Stderr: cb2bde55f71f Extracting [==============> ] 14.68MB/50.34MB Stderr: cb2bde55f71f Extracting [======================> ] 22.54MB/50.34MB Stderr: cb2bde55f71f Extracting [==============================> ] 30.93MB/50.34MB Stderr: cb2bde55f71f Extracting [=========================================> ] 41.42MB/50.34MB Stderr: cb2bde55f71f Extracting [==============================================> ] 46.66MB/50.34MB Stderr: cb2bde55f71f Extracting [===============================================> ] 47.71MB/50.34MB Stderr: cb2bde55f71f Extracting [===============================================> ] 48.23MB/50.34MB Stderr: cb2bde55f71f Extracting [================================================> ] 48.76MB/50.34MB Stderr: cb2bde55f71f Extracting [================================================> ] 49.28MB/50.34MB Stderr: cb2bde55f71f Extracting [=================================================> ] 50.33MB/50.34MB Stderr: cb2bde55f71f Extracting [==================================================>] 50.34MB/50.34MB Stderr: cb2bde55f71f Pull complete Stderr: 9d0e0719fbe0 Extracting [=> ] 32.77kB/1.261MB Stderr: 9d0e0719fbe0 Extracting [==================================================>] 1.261MB/1.261MB Stderr: 9d0e0719fbe0 Extracting [==================================================>] 1.261MB/1.261MB Stderr: 9d0e0719fbe0 Pull complete Stderr: 6f063dbd7a5d Downloading [==================================================>] 446B/446B Stderr: 6f063dbd7a5d Download complete Stderr: 6f063dbd7a5d Extracting [==================================================>] 446B/446B Stderr: 6f063dbd7a5d Extracting [==================================================>] 446B/446B Stderr: a39ef2f62dc8 Downloading [> ] 3.29kB/209.4kB Stderr: f9e3e3d8f042 Downloading [==================================================>] 116B/116B Stderr: f9e3e3d8f042 Download complete Stderr: 6f063dbd7a5d Pull complete Stderr: f9e3e3d8f042 Extracting [==================================================>] 116B/116B Stderr: f9e3e3d8f042 Extracting [==================================================>] 116B/116B Stderr: f9e3e3d8f042 Pull complete Stderr: a39ef2f62dc8 Downloading [==================================================>] 209.4kB/209.4kB Stderr: a39ef2f62dc8 Verifying Checksum Stderr: a39ef2f62dc8 Download complete Stderr: a39ef2f62dc8 Extracting [=======> ] 32.77kB/209.4kB Stderr: a39ef2f62dc8 Extracting [==================================================>] 209.4kB/209.4kB Stderr: a39ef2f62dc8 Extracting [==================================================>] 209.4kB/209.4kB Stderr: a39ef2f62dc8 Pull complete Stderr: 9a21c6b23f0e Downloading [> ] 15.58kB/794kB Stderr: efeb7b313b67 Downloading [=> ] 15.58kB/458.8kB Stderr: 6fef65209747 Downloading [> ] 375.1kB/36.34MB Stderr: 9a21c6b23f0e Downloading [=====================> ] 342.3kB/794kB Stderr: efeb7b313b67 Downloading [=================================> ] 309.5kB/458.8kB Stderr: efeb7b313b67 Verifying Checksum Stderr: efeb7b313b67 Download complete Stderr: 9a21c6b23f0e Downloading [==================================================>] 794kB/794kB Stderr: 9a21c6b23f0e Download complete Stderr: 9a21c6b23f0e Extracting [==> ] 32.77kB/794kB Stderr: 6fef65209747 Downloading [===> ] 2.259MB/36.34MB Stderr: 9a21c6b23f0e Extracting [============================> ] 458.8kB/794kB Stderr: 6fef65209747 Downloading [========> ] 6.404MB/36.34MB Stderr: 9a21c6b23f0e Extracting [=============================================> ] 720.9kB/794kB Stderr: 9a21c6b23f0e Extracting [==================================================>] 794kB/794kB Stderr: 9a21c6b23f0e Pull complete Stderr: efeb7b313b67 Extracting [===> ] 32.77kB/458.8kB Stderr: 6fef65209747 Downloading [====================> ] 14.7MB/36.34MB Stderr: efeb7b313b67 Extracting [==================================================>] 458.8kB/458.8kB Stderr: efeb7b313b67 Extracting [==================================================>] 458.8kB/458.8kB Stderr: 6fef65209747 Downloading [====================================> ] 26.76MB/36.34MB Stderr: efeb7b313b67 Pull complete Stderr: 6fef65209747 Verifying Checksum Stderr: 6fef65209747 Download complete Stderr: 6fef65209747 Extracting [> ] 393.2kB/36.34MB Stderr: 6fef65209747 Extracting [=> ] 1.18MB/36.34MB Stderr: 6fef65209747 Extracting [==> ] 1.573MB/36.34MB Stderr: 6fef65209747 Extracting [==> ] 1.966MB/36.34MB Stderr: 6fef65209747 Extracting [===> ] 2.753MB/36.34MB Stderr: 6fef65209747 Extracting [=====> ] 3.932MB/36.34MB Stderr: 6fef65209747 Extracting [======> ] 4.719MB/36.34MB Stderr: 6fef65209747 Extracting [========> ] 5.898MB/36.34MB Stderr: 6fef65209747 Extracting [========> ] 6.291MB/36.34MB Stderr: 6fef65209747 Extracting [=========> ] 7.078MB/36.34MB Stderr: 6fef65209747 Extracting [==========> ] 7.864MB/36.34MB Stderr: 6fef65209747 Extracting [===========> ] 8.258MB/36.34MB Stderr: 6fef65209747 Extracting [===========> ] 8.651MB/36.34MB Stderr: 6fef65209747 Extracting [=============> ] 9.83MB/36.34MB Stderr: 6fef65209747 Extracting [=================> ] 12.98MB/36.34MB Stderr: 6fef65209747 Extracting [==================> ] 13.76MB/36.34MB Stderr: 6fef65209747 Extracting [====================> ] 14.94MB/36.34MB Stderr: 6fef65209747 Extracting [=====================> ] 15.73MB/36.34MB Stderr: 6fef65209747 Extracting [======================> ] 16.12MB/36.34MB Stderr: 6fef65209747 Extracting [=======================> ] 17.3MB/36.34MB Stderr: 6fef65209747 Extracting [============================> ] 20.45MB/36.34MB Stderr: 6fef65209747 Extracting [======================================> ] 28.31MB/36.34MB Stderr: 6fef65209747 Extracting [==========================================> ] 31.06MB/36.34MB Stderr: 6fef65209747 Extracting [==================================================>] 36.34MB/36.34MB Stderr: 6fef65209747 Pull complete Stderr: 3d377e512a83 Downloading [==================================================>] 2.862kB/2.862kB Stderr: 3d377e512a83 Download complete Stderr: 3d377e512a83 Extracting [==================================================>] 2.862kB/2.862kB Stderr: 3d377e512a83 Extracting [==================================================>] 2.862kB/2.862kB Stderr: 3d377e512a83 Pull complete Stderr: azurite1 Pulled Setup ZooKeeper Creating internal ZooKeeper dirs: ['/ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw8/keeper1/log', '/ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw8/keeper1/config', '/ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw8/keeper1/coordination', '/ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw8/keeper2/log', '/ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw8/keeper2/config', '/ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw8/keeper2/coordination', '/ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw8/keeper3/log', '/ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw8/keeper3/config', '/ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw8/keeper3/coordination'] Command:[docker compose --project-name roottestazureblobstoragezerocopyreplication-gw8 --env-file /ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw8/.env --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --verbose up -d] Executing query select dictGetOrDefault('LocalClickHouse_direct_', 'String_', toUInt64(9999999), toString('hi')) on local_node Executing query BACKUP DATABASE mydb ON CLUSTER 'cluster' TO Disk('backups', '29') on node2 Executing query select 20 on node1 Executing query select dictGetInt32OrDefault('MySQL_direct_', 'Int32_', toUInt64(9999999), toInt32(-77)) on node Executing query select dictGetStringOrDefault('LocalClickHouse_direct_', 'String_', toUInt64(9999999), toString('hi')) on local_node Executing query select dictGet('MySQL_direct_', 'Int64_', toUInt64(1)) on node Stderr:time="2025-04-02T02:21:41Z" level=trace msg="Docker Desktop integration not enabled" Stderr: Network roottestazureblobstoragezerocopyreplication-gw8_default Creating Stderr: Network roottestazureblobstoragezerocopyreplication-gw8_default Created Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-zoo3-1 Creating Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-zoo1-1 Creating Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-zoo2-1 Creating Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-zoo1-1 Created Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-zoo2-1 Created Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-zoo3-1 Created Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-zoo2-1 Starting Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-zoo3-1 Starting Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-zoo1-1 Starting Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-zoo3-1 Started Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-zoo1-1 Started Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-zoo2-1 Started Stderr:time="2025-04-02T02:21:41Z" level=debug msg="otel error" error="" Stderr:time="2025-04-02T02:21:41Z" level=debug msg="otel error" error="" Wait ZooKeeper to start get_instance_ip instance_name=zoo1 http://localhost:None "GET /v1.46/containers/roottestazureblobstoragezerocopyreplication-gw8-zoo1-1/json HTTP/1.1" 200 None get_kazoo_client: zoo1, ip:172.16.1.4, port:2181, use_ssl:False Connecting to 172.16.1.4(172.16.1.4):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Executing query select dictGet('LocalClickHouse_direct_', 'Float32_', toUInt64(1)) on local_node Executing query select 20 on node1 Connecting to 172.16.1.4(172.16.1.4):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Executing query DROP DATABASE mydb ON CLUSTER 'cluster' SYNC on node1 Executing query select dictGetInt64('MySQL_direct_', 'Int64_', toUInt64(1)) on node Connecting to 172.16.1.4(172.16.1.4):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Executing query select dictGetFloat32('LocalClickHouse_direct_', 'Float32_', toUInt64(1)) on local_node Executing query select dictHas('MySQL_direct_', toUInt64(1)) on node Executing query RESTORE DATABASE mydb ON CLUSTER 'cluster' FROM Disk('backups', '29') SETTINGS replica_num_in_backup=1 on node1 Executing query select dictHas('LocalClickHouse_direct_', toUInt64(1)) on local_node Executing query select dictGetOrDefault('MySQL_direct_', 'Int64_', toUInt64(9999999), toInt64(-88)) on node Executing query select 20 on node1 Connecting to 172.16.1.4(172.16.1.4):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Executing query select dictGetInt64OrDefault('MySQL_direct_', 'Int64_', toUInt64(9999999), toInt64(-88)) on node Executing query select dictGetOrDefault('LocalClickHouse_direct_', 'Float32_', toUInt64(9999999), toFloat32(555.11)) on local_node [gw7] PASSED test_async_load_databases/test.py::test_dict_get_data test_async_load_databases/test.py::test_multiple_tables Executing query create table test.table_0 (n UInt64, s String) engine=MergeTree order by n as select number, randomString(100) from numbers(100) on node1 Executing query select dictGet('MySQL_direct_', 'UUID_', toUInt64(1)) on node Executing query SELECT name, type FROM system.columns WHERE database='mydb' AND table='tbl' on node1 Executing query select dictGetFloat32OrDefault('LocalClickHouse_direct_', 'Float32_', toUInt64(9999999), toFloat32(555.11)) on local_node Executing query create table test.table_1 (n UInt64, s String) engine=MergeTree order by n as select number, randomString(100) from numbers(100) on node1 Executing query select dictGetUUID('MySQL_direct_', 'UUID_', toUInt64(1)) on node Executing query SELECT name, type FROM system.columns WHERE database='mydb' AND table='tbl' on node2 Executing query select dictGet('LocalClickHouse_direct_', 'Float64_', toUInt64(1)) on local_node Executing query create table test.table_2 (n UInt64, s String) engine=MergeTree order by n as select number, randomString(100) from numbers(100) on node1 Executing query select dictHas('MySQL_direct_', toUInt64(1)) on node Executing query DROP DATABASE mydb ON CLUSTER 'cluster' SYNC on node1 Executing query select dictGetFloat64('LocalClickHouse_direct_', 'Float64_', toUInt64(1)) on local_node Executing query create table test.table_3 (n UInt64, s String) engine=MergeTree order by n as select number, randomString(100) from numbers(100) on node1 Connecting to 172.16.1.4(172.16.1.4):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Executing query select dictGetOrDefault('MySQL_direct_', 'UUID_', toUInt64(9999999), toUUID('550e8400-0000-0000-0000-000000000000')) on node Executing query select dictHas('LocalClickHouse_direct_', toUInt64(1)) on local_node Executing query create table test.table_4 (n UInt64, s String) engine=MergeTree order by n as select number, randomString(100) from numbers(100) on node1 Executing query RESTORE DATABASE mydb ON CLUSTER 'cluster' FROM Disk('backups', '29') SETTINGS replica_num_in_backup=2 on node2 Executing query select dictGetUUIDOrDefault('MySQL_direct_', 'UUID_', toUInt64(9999999), toUUID('550e8400-0000-0000-0000-000000000000')) on node Stderr: Container roottestdictionariesredis-gw0-node-1 Stopping Stderr: Container roottestdictionariesredis-gw0-redis1-1 Stopping Stderr: Container roottestdictionariesredis-gw0-redis1-1 Stopped Stderr: Container roottestdictionariesredis-gw0-node-1 Stopped Command:[bash -c [ -f /ClickHouse/tests/integration/test_dictionaries_redis/_instances-0-gw0/node/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_dictionaries_redis/_instances-0-gw0/node/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[docker compose --env-file /ClickHouse/tests/integration/test_dictionaries_redis/_instances-0-gw0/.env --project-name roottestdictionariesredis-gw0 --file /ClickHouse/tests/integration/test_dictionaries_redis/_instances-0-gw0/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_redis.yml down --volumes] Executing query select dictGetOrDefault('LocalClickHouse_direct_', 'Float64_', toUInt64(9999999), toFloat64(777.11)) on local_node Executing query create table test.table_5 (n UInt64, s String) engine=MergeTree order by n as select number, randomString(100) from numbers(100) on node1 Executing query select dictGet('MySQL_direct_', 'Date_', toUInt64(1)) on node Executing query select dictGetFloat64OrDefault('LocalClickHouse_direct_', 'Float64_', toUInt64(9999999), toFloat64(777.11)) on local_node Executing query create table test.table_6 (n UInt64, s String) engine=MergeTree order by n as select number, randomString(100) from numbers(100) on node1 Executing query select dictGetDate('MySQL_direct_', 'Date_', toUInt64(1)) on node Executing query SELECT name, type FROM system.columns WHERE database='mydb' AND table='tbl' on node1 Stderr: Container roottestdictionariesredis-gw0-redis1-1 Stopping Stderr: Container roottestdictionariesredis-gw0-node-1 Stopping Stderr: Container roottestdictionariesredis-gw0-redis1-1 Stopped Stderr: Container roottestdictionariesredis-gw0-redis1-1 Removing Stderr: Container roottestdictionariesredis-gw0-node-1 Stopped Stderr: Container roottestdictionariesredis-gw0-node-1 Removing Stderr: Container roottestdictionariesredis-gw0-node-1 Removed Stderr: Container roottestdictionariesredis-gw0-redis1-1 Removed Stderr: Network roottestdictionariesredis-gw0_default Removing Stderr: Network roottestdictionariesredis-gw0_default Removed Cleanup called Docker networks for project roottestdictionariesredis-gw0 are NETWORK ID NAME DRIVER SCOPE Executing query select dictGet('LocalClickHouse_direct_', 'ParentKeyField', toUInt64(1)) on local_node Docker containers for project roottestdictionariesredis-gw0 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestdictionariesredis-gw0 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestdictionariesredis-gw0-.*-1$' --format '{{.ID}}:{{.Names}}'] Executing query select dictHas('MySQL_direct_', toUInt64(1)) on node Unstopped containers: {} No running containers for project: roottestdictionariesredis-gw0 Trying to prune unused networks... Executing query create table test.table_7 (n UInt64, s String) engine=MergeTree order by n as select number, randomString(100) from numbers(100) on node1 Trying to prune unused images... Command:[docker image prune -f] Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Stdout:2 Command:[docker volume prune -f] Stdout:Total reclaimed space: 0B Volumes pruned: 2 Executing query SELECT name, type FROM system.columns WHERE database='mydb' AND table='tbl' on node2 Executing query select dictGetOrDefault('MySQL_direct_', 'Date_', toUInt64(9999999), toDate('2018-12-30')) on node Executing query select dictGetUInt64('LocalClickHouse_direct_', 'ParentKeyField', toUInt64(1)) on local_node Executing query create table test.table_8 (n UInt64, s String) engine=MergeTree order by n as select number, randomString(100) from numbers(100) on node1 Executing query DROP TABLE IF EXISTS tbl ON CLUSTER 'cluster3' SYNC on node1 [gw2] PASSED test_backup_restore_on_cluster/test.py::test_table_in_replicated_database_with_not_synced_def Executing query select dictGetDateOrDefault('MySQL_direct_', 'Date_', toUInt64(9999999), toDate('2018-12-30')) on node Executing query select dictHas('LocalClickHouse_direct_', toUInt64(1)) on local_node Executing query create table test.table_9 (n UInt64, s String) engine=MergeTree order by n as select number, randomString(100) from numbers(100) on node1 Connecting to 172.16.1.4(172.16.1.4):2181, use_ssl: False Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Executing query select dictGet('MySQL_direct_', 'DateTime_', toUInt64(1)) on node Zookeeper connection established, state: CONNECTED Sending request(xid=1): GetChildren(path='/', watcher=None) Received response(xid=1): ['keeper'] Sending request(xid=2): Close() Connection dropped: socket connection broken Transition to CONNECTING Zookeeper connection lost Executing query DROP TABLE IF EXISTS tbl2 ON CLUSTER 'cluster3' SYNC on node1 Executing query select dictGetOrDefault('LocalClickHouse_direct_', 'ParentKeyField', toUInt64(9999999), toUInt64(444)) on local_node Executing query create table test.table_10 (n UInt64, s String) engine=MergeTree order by n as select number, randomString(100) from numbers(100) on node1 Failed connecting to Zookeeper within the connection retry policy. Zookeeper session closed, state: CLOSED get_instance_ip instance_name=zoo2 http://localhost:None "GET /v1.46/containers/roottestazureblobstoragezerocopyreplication-gw8-zoo2-1/json HTTP/1.1" 200 None get_kazoo_client: zoo2, ip:172.16.1.2, port:2181, use_ssl:False Connecting to 172.16.1.2(172.16.1.2):2181, use_ssl: False Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Zookeeper connection established, state: CONNECTED Sending request(xid=1): GetChildren(path='/', watcher=None) Received response(xid=1): ['keeper'] Sending request(xid=2): Close() Connection dropped: socket connection broken Transition to CONNECTING Zookeeper connection lost Failed connecting to Zookeeper within the connection retry policy. Zookeeper session closed, state: CLOSED get_instance_ip instance_name=zoo3 http://localhost:None "GET /v1.46/containers/roottestazureblobstoragezerocopyreplication-gw8-zoo3-1/json HTTP/1.1" 200 None get_kazoo_client: zoo3, ip:172.16.1.3, port:2181, use_ssl:False Connecting to 172.16.1.3(172.16.1.3):2181, use_ssl: False Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Zookeeper connection established, state: CONNECTED Sending request(xid=1): GetChildren(path='/', watcher=None) Received response(xid=1): ['keeper'] Sending request(xid=2): Close() Connection dropped: socket connection broken Transition to CONNECTING Zookeeper connection lost Executing query select dictGetDateTime('MySQL_direct_', 'DateTime_', toUInt64(1)) on node Failed connecting to Zookeeper within the connection retry policy. Zookeeper session closed, state: CLOSED All instances of ZooKeeper started: ('zoo1', 'zoo2', 'zoo3') Trying to create Azurite instance by command docker compose --project-name roottestazureblobstoragezerocopyreplication-gw8 --env-file /ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw8/.env --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_azurite.yml --verbose up -d Command:[docker compose --project-name roottestazureblobstoragezerocopyreplication-gw8 --env-file /ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw8/.env --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_azurite.yml --verbose up -d] Executing query select dictGetUInt64OrDefault('LocalClickHouse_direct_', 'ParentKeyField', toUInt64(9999999), toUInt64(444)) on local_node Executing query create table test.table_11 (n UInt64, s String) engine=MergeTree order by n as select number, randomString(100) from numbers(100) on node1 Executing query DROP DATABASE IF EXISTS mydb ON CLUSTER 'cluster3' SYNC on node1 Executing query select dictHas('MySQL_direct_', toUInt64(1)) on node Executing query select dictGet('LocalClickHouse_direct_', 'UInt8_', toUInt64(2)) on local_node Executing query create table test.table_12 (n UInt64, s String) engine=MergeTree order by n as select number, randomString(100) from numbers(100) on node1 Stderr:time="2025-04-02T02:21:46Z" level=trace msg="Docker Desktop integration not enabled" Stderr: Volume "roottestazureblobstoragezerocopyreplication-gw8_data1-1" Creating Stderr: Volume "roottestazureblobstoragezerocopyreplication-gw8_data1-1" Created Stderr:time="2025-04-02T02:21:46Z" level=warning msg="Found orphan containers ([roottestazureblobstoragezerocopyreplication-gw8-zoo1-1 roottestazureblobstoragezerocopyreplication-gw8-zoo3-1 roottestazureblobstoragezerocopyreplication-gw8-zoo2-1]) for this project. If you removed or renamed this service in your compose file, you can run this command with the --remove-orphans flag to clean it up." Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-azurite1-1 Creating Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-azurite1-1 Created Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-azurite1-1 Starting Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-azurite1-1 Started Stderr:time="2025-04-02T02:21:46Z" level=debug msg="otel error" error="" Stderr:time="2025-04-02T02:21:46Z" level=debug msg="otel error" error="" Trying to connect to Azurite Executing query DROP DATABASE IF EXISTS mydb2 ON CLUSTER 'cluster3' SYNC on node1 Executing query select dictGetOrDefault('MySQL_direct_', 'DateTime_', toUInt64(9999999), toDateTime('2018-12-30 00:00:00')) on node Executing query select dictGetUInt8('LocalClickHouse_direct_', 'UInt8_', toUInt64(2)) on local_node Executing query create table test.table_13 (n UInt64, s String) engine=MergeTree order by n as select number, randomString(100) from numbers(100) on node1 Executing query DROP USER IF EXISTS u1, u2 ON CLUSTER 'cluster3' on node1 Executing query select dictGetDateTimeOrDefault('MySQL_direct_', 'DateTime_', toUInt64(9999999), toDateTime('2018-12-30 00:00:00')) on node Executing query select dictHas('LocalClickHouse_direct_', toUInt64(2)) on local_node Executing query create table test.table_14 (n UInt64, s String) engine=MergeTree order by n as select number, randomString(100) from numbers(100) on node1 Executing query select dictGet('MySQL_direct_', 'String_', toUInt64(1)) on node test_backup_restore_on_cluster/test.py::test_table_with_parts_in_queue_considered_non_empty Executing query CREATE DATABASE mydb ON CLUSTER 'cluster' ENGINE=Replicated('/clickhouse/path/','{shard}','{replica}') on node1 Executing query select dictGetOrDefault('LocalClickHouse_direct_', 'UInt8_', toUInt64(9999999), toUInt8(55)) on local_node Executing query create table test.table_15 (n UInt64, s String) engine=MergeTree order by n as select number, randomString(100) from numbers(100) on node1 Request URL: 'http://127.0.0.1:30400/devstoreaccount1/?restype=REDACTED&comp=REDACTED' Request method: 'GET' Request headers: 'x-ms-version': 'REDACTED' 'Accept': 'application/xml' 'User-Agent': 'azsdk-python-storage-blob/12.19.0 Python/3.10.12 (Linux-5.15.0-130-generic-x86_64-with-glibc2.35)' 'x-ms-date': 'REDACTED' 'x-ms-client-request-id': '3a7955dc-0f69-11f0-bbd2-0242ac110002' 'Authorization': 'REDACTED' No body was attached to the request Starting new HTTP connection (1): 127.0.0.1:30400 Executing query select dictGetString('MySQL_direct_', 'String_', toUInt64(1)) on node http://127.0.0.1:30400 "GET /devstoreaccount1/?restype=account&comp=properties HTTP/1.1" 200 0 Response status: 200 Response headers: 'Server': 'Azurite-Blob/3.34.0' 'x-ms-client-request-id': '3a7955dc-0f69-11f0-bbd2-0242ac110002' 'x-ms-request-id': 'f841a5b6-263d-4d4d-b01a-9dca7b6dac3a' 'x-ms-version': 'REDACTED' 'date': 'Wed, 02 Apr 2025 02:21:47 GMT' 'x-ms-sku-name': 'REDACTED' 'x-ms-account-kind': 'REDACTED' 'x-ms-is-hns-enabled': 'REDACTED' 'Connection': 'keep-alive' 'Keep-Alive': 'REDACTED' 'Content-Length': '0' {'client_request_id': '3a7955dc-0f69-11f0-bbd2-0242ac110002', 'request_id': 'f841a5b6-263d-4d4d-b01a-9dca7b6dac3a', 'version': '2025-05-05', 'date': datetime.datetime(2025, 4, 2, 2, 21, 47, tzinfo=datetime.timezone.utc), 'sku_name': 'Standard_RAGRS', 'account_kind': 'StorageV2', 'is_hns_enabled': False} Request URL: 'http://127.0.0.1:30400/devstoreaccount1/?comp=REDACTED&prefix=REDACTED&include=REDACTED' Request method: 'GET' Request headers: 'x-ms-version': 'REDACTED' 'Accept': 'application/xml' 'User-Agent': 'azsdk-python-storage-blob/12.19.0 Python/3.10.12 (Linux-5.15.0-130-generic-x86_64-with-glibc2.35)' 'x-ms-date': 'REDACTED' 'x-ms-client-request-id': '3a7f3b0a-0f69-11f0-bbd2-0242ac110002' 'Authorization': 'REDACTED' No body was attached to the request http://127.0.0.1:30400 "GET /devstoreaccount1/?comp=list&prefix=azurite-container&include= HTTP/1.1" 200 None Response status: 200 Response headers: 'Server': 'Azurite-Blob/3.34.0' 'x-ms-client-request-id': '3a7f3b0a-0f69-11f0-bbd2-0242ac110002' 'x-ms-request-id': '529b2d21-78a2-4d07-9db0-1d4ee656709f' 'x-ms-version': 'REDACTED' 'content-type': 'application/xml' 'Date': 'Wed, 02 Apr 2025 02:21:47 GMT' 'Connection': 'keep-alive' 'Keep-Alive': 'REDACTED' 'Transfer-Encoding': 'chunked' Request URL: 'http://127.0.0.1:30400/devstoreaccount1/azurite-container?restype=REDACTED' Request method: 'GET' Request headers: 'x-ms-version': 'REDACTED' 'Accept': 'application/xml' 'User-Agent': 'azsdk-python-storage-blob/12.19.0 Python/3.10.12 (Linux-5.15.0-130-generic-x86_64-with-glibc2.35)' 'x-ms-date': 'REDACTED' 'x-ms-client-request-id': '3a818964-0f69-11f0-bbd2-0242ac110002' 'Authorization': 'REDACTED' No body was attached to the request http://127.0.0.1:30400 "GET /devstoreaccount1/azurite-container?restype=container HTTP/1.1" 404 None Response status: 404 Response headers: 'Server': 'Azurite-Blob/3.34.0' 'x-ms-error-code': 'ContainerNotFound' 'x-ms-request-id': 'ea708c04-dabb-4cba-ab37-31f5bce59074' 'content-type': 'application/xml' 'Date': 'Wed, 02 Apr 2025 02:21:47 GMT' 'Connection': 'keep-alive' 'Keep-Alive': 'REDACTED' 'Transfer-Encoding': 'chunked' azurite container 'azurite-container' doesn't exist, creating it Request URL: 'http://127.0.0.1:30400/devstoreaccount1/azurite-container?restype=REDACTED' Request method: 'PUT' Request headers: 'x-ms-version': 'REDACTED' 'Accept': 'application/xml' 'User-Agent': 'azsdk-python-storage-blob/12.19.0 Python/3.10.12 (Linux-5.15.0-130-generic-x86_64-with-glibc2.35)' 'x-ms-date': 'REDACTED' 'x-ms-client-request-id': '3a831d74-0f69-11f0-bbd2-0242ac110002' 'Authorization': 'REDACTED' No body was attached to the request http://127.0.0.1:30400 "PUT /devstoreaccount1/azurite-container?restype=container HTTP/1.1" 201 0 Response status: 201 Response headers: 'Server': 'Azurite-Blob/3.34.0' 'etag': '"0x200613B17F482C0"' 'last-modified': 'Wed, 02 Apr 2025 02:21:47 GMT' 'x-ms-client-request-id': '3a831d74-0f69-11f0-bbd2-0242ac110002' 'x-ms-request-id': '2ca45a07-2e9c-4be5-963c-94189d7a0a78' 'x-ms-version': 'REDACTED' 'Date': 'Wed, 02 Apr 2025 02:21:47 GMT' 'Connection': 'keep-alive' 'Keep-Alive': 'REDACTED' 'Content-Length': '0' ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw8/.env --project-name roottestazureblobstoragezerocopyreplication-gw8 --file /ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw8/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_azurite.yml --file /ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw8/node2/docker-compose.yml up -d --no-recreate') Command:[docker compose --env-file /ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw8/.env --project-name roottestazureblobstoragezerocopyreplication-gw8 --file /ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw8/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_azurite.yml --file /ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw8/node2/docker-compose.yml up -d --no-recreate] Executing query CREATE TABLE mydb.test (`x` UInt32) ENGINE = ReplicatedMergeTree ORDER BY x on node1 Executing query create table test.table_16 (n UInt64, s String) engine=MergeTree order by n as select number, randomString(100) from numbers(100) on node1 Executing query select dictGetUInt8OrDefault('LocalClickHouse_direct_', 'UInt8_', toUInt64(9999999), toUInt8(55)) on local_node Executing query select dictHas('MySQL_direct_', toUInt64(1)) on node Executing query create table test.table_17 (n UInt64, s String) engine=MergeTree order by n as select number, randomString(100) from numbers(100) on node1 Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-zoo2-1 Running Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-azurite1-1 Running Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-zoo3-1 Running Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-zoo1-1 Running Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-node2-1 Creating Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-node1-1 Creating Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-node2-1 Created Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-node1-1 Created Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-node2-1 Starting Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-node1-1 Starting Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-node1-1 Started Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-node2-1 Started ClickHouse instance created get_instance_ip instance_name=node1 http://localhost:None "GET /v1.46/containers/roottestazureblobstoragezerocopyreplication-gw8-node1-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node1 http://localhost:None "GET /v1.46/containers/roottestazureblobstoragezerocopyreplication-gw8-node1-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node1, ip: 172.16.1.7... http://localhost:None "GET /v1.46/containers/roottestazureblobstoragezerocopyreplication-gw8-node1-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/57432c314649355826adf676e3f4f07a3eb5e8f6400fc8d5fcafdb9d3deceab3/json HTTP/1.1" 200 None Executing query INSERT INTO mydb.test SELECT number AS x FROM numbers(10000000) on node1 http://localhost:None "GET /v1.46/containers/57432c314649355826adf676e3f4f07a3eb5e8f6400fc8d5fcafdb9d3deceab3/json HTTP/1.1" 200 None Executing query select dictGetOrDefault('MySQL_direct_', 'String_', toUInt64(9999999), toString('hi')) on node Executing query select dictGet('LocalClickHouse_direct_', 'UInt16_', toUInt64(2)) on local_node http://localhost:None "GET /v1.46/containers/57432c314649355826adf676e3f4f07a3eb5e8f6400fc8d5fcafdb9d3deceab3/json HTTP/1.1" 200 None Executing query create table test.table_18 (n UInt64, s String) engine=MergeTree order by n as select number, randomString(100) from numbers(100) on node1 http://localhost:None "GET /v1.46/containers/57432c314649355826adf676e3f4f07a3eb5e8f6400fc8d5fcafdb9d3deceab3/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/57432c314649355826adf676e3f4f07a3eb5e8f6400fc8d5fcafdb9d3deceab3/json HTTP/1.1" 200 None Executing query select dictGetStringOrDefault('MySQL_direct_', 'String_', toUInt64(9999999), toString('hi')) on node Executing query select dictGetUInt16('LocalClickHouse_direct_', 'UInt16_', toUInt64(2)) on local_node http://localhost:None "GET /v1.46/containers/57432c314649355826adf676e3f4f07a3eb5e8f6400fc8d5fcafdb9d3deceab3/json HTTP/1.1" 200 None Executing query create table test.table_19 (n UInt64, s String) engine=MergeTree order by n as select number, randomString(100) from numbers(100) on node1 http://localhost:None "GET /v1.46/containers/57432c314649355826adf676e3f4f07a3eb5e8f6400fc8d5fcafdb9d3deceab3/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/57432c314649355826adf676e3f4f07a3eb5e8f6400fc8d5fcafdb9d3deceab3/json HTTP/1.1" 200 None Executing query select dictHas('LocalClickHouse_direct_', toUInt64(2)) on local_node Executing query select dictGet('MySQL_direct_', 'Float32_', toUInt64(1)) on node http://localhost:None "GET /v1.46/containers/57432c314649355826adf676e3f4f07a3eb5e8f6400fc8d5fcafdb9d3deceab3/json HTTP/1.1" 200 None run container_id:roottestasyncloaddatabases-gw7-node1-1 detach:False nothrow:True cmd: ['bash', '-c', 'ps -C clickhouse'] Command:[docker exec -u root roottestasyncloaddatabases-gw7-node1-1 bash -c ps -C clickhouse] Stdout: PID TTY TIME CMD Stdout: 1581 ? 00:00:13 clickhouse run container_id:roottestasyncloaddatabases-gw7-node1-1 detach:False nothrow:False cmd: ['bash', '-c', 'pkill clickhouse'] Command:[docker exec -u root roottestasyncloaddatabases-gw7-node1-1 bash -c pkill clickhouse] http://localhost:None "GET /v1.46/containers/57432c314649355826adf676e3f4f07a3eb5e8f6400fc8d5fcafdb9d3deceab3/json HTTP/1.1" 200 None run container_id:roottestasyncloaddatabases-gw7-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw7-node1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select dictGetFloat32('MySQL_direct_', 'Float32_', toUInt64(1)) on node Stdout:1581 http://localhost:None "GET /v1.46/containers/57432c314649355826adf676e3f4f07a3eb5e8f6400fc8d5fcafdb9d3deceab3/json HTTP/1.1" 200 None Executing query select dictGetOrDefault('LocalClickHouse_direct_', 'UInt16_', toUInt64(9999999), toUInt16(66)) on local_node http://localhost:None "GET /v1.46/containers/57432c314649355826adf676e3f4f07a3eb5e8f6400fc8d5fcafdb9d3deceab3/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/57432c314649355826adf676e3f4f07a3eb5e8f6400fc8d5fcafdb9d3deceab3/json HTTP/1.1" 200 None Executing query select dictHas('MySQL_direct_', toUInt64(1)) on node http://localhost:None "GET /v1.46/containers/57432c314649355826adf676e3f4f07a3eb5e8f6400fc8d5fcafdb9d3deceab3/json HTTP/1.1" 200 None Executing query select dictGetUInt16OrDefault('LocalClickHouse_direct_', 'UInt16_', toUInt64(9999999), toUInt16(66)) on local_node http://localhost:None "GET /v1.46/containers/57432c314649355826adf676e3f4f07a3eb5e8f6400fc8d5fcafdb9d3deceab3/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/57432c314649355826adf676e3f4f07a3eb5e8f6400fc8d5fcafdb9d3deceab3/json HTTP/1.1" 200 None Executing query select dictGetOrDefault('MySQL_direct_', 'Float32_', toUInt64(9999999), toFloat32(555.11)) on node http://localhost:None "GET /v1.46/containers/57432c314649355826adf676e3f4f07a3eb5e8f6400fc8d5fcafdb9d3deceab3/json HTTP/1.1" 200 None Executing query select dictGet('LocalClickHouse_direct_', 'UInt32_', toUInt64(2)) on local_node http://localhost:None "GET /v1.46/containers/57432c314649355826adf676e3f4f07a3eb5e8f6400fc8d5fcafdb9d3deceab3/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/57432c314649355826adf676e3f4f07a3eb5e8f6400fc8d5fcafdb9d3deceab3/json HTTP/1.1" 200 None Executing query select dictGetFloat32OrDefault('MySQL_direct_', 'Float32_', toUInt64(9999999), toFloat32(555.11)) on node http://localhost:None "GET /v1.46/containers/57432c314649355826adf676e3f4f07a3eb5e8f6400fc8d5fcafdb9d3deceab3/json HTTP/1.1" 200 None Executing query select dictGetUInt32('LocalClickHouse_direct_', 'UInt32_', toUInt64(2)) on local_node run container_id:roottestasyncloaddatabases-gw7-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw7-node1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] http://localhost:None "GET /v1.46/containers/57432c314649355826adf676e3f4f07a3eb5e8f6400fc8d5fcafdb9d3deceab3/json HTTP/1.1" 200 None Stdout:1581 Executing query BACKUP DATABASE mydb TO Disk('backups', '30') on node1 http://localhost:None "GET /v1.46/containers/57432c314649355826adf676e3f4f07a3eb5e8f6400fc8d5fcafdb9d3deceab3/json HTTP/1.1" 200 None Executing query select dictGet('MySQL_direct_', 'Float64_', toUInt64(1)) on node http://localhost:None "GET /v1.46/containers/57432c314649355826adf676e3f4f07a3eb5e8f6400fc8d5fcafdb9d3deceab3/json HTTP/1.1" 200 None Executing query select dictHas('LocalClickHouse_direct_', toUInt64(2)) on local_node http://localhost:None "GET /v1.46/containers/57432c314649355826adf676e3f4f07a3eb5e8f6400fc8d5fcafdb9d3deceab3/json HTTP/1.1" 200 None ClickHouse node1 started get_instance_ip instance_name=node2 http://localhost:None "GET /v1.46/containers/roottestazureblobstoragezerocopyreplication-gw8-node2-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node2 http://localhost:None "GET /v1.46/containers/roottestazureblobstoragezerocopyreplication-gw8-node2-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node2, ip: 172.16.1.6... http://localhost:None "GET /v1.46/containers/roottestazureblobstoragezerocopyreplication-gw8-node2-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/516a6556f56d885b246599f111e3a11415ea3eb6fbe196a6f96e18227eaaa153/json HTTP/1.1" 200 None ClickHouse node2 started Cluster started Executing query CREATE TABLE blob_storage_table ON CLUSTER test_cluster ( id Int64, data String ) ENGINE=ReplicatedMergeTree('/clickhouse/tables/blob_storage_table', '{replica}') ORDER BY id SETTINGS storage_policy='blob_storage_policy',old_parts_lifetime=1 on node1 Executing query select dictGetFloat64('MySQL_direct_', 'Float64_', toUInt64(1)) on node Executing query select dictGetOrDefault('LocalClickHouse_direct_', 'UInt32_', toUInt64(9999999), toUInt32(77)) on local_node Executing query DROP DATABASE mydb SYNC on node1 Executing query select dictHas('MySQL_direct_', toUInt64(1)) on node Executing query SELECT COUNT(*) FROM blob_storage_table FORMAT Values on node1 Executing query select dictGetUInt32OrDefault('LocalClickHouse_direct_', 'UInt32_', toUInt64(9999999), toUInt32(77)) on local_node Executing query RESTORE DATABASE mydb FROM Disk('backups', '30') on node1 Executing query select dictGetOrDefault('MySQL_direct_', 'Float64_', toUInt64(9999999), toFloat64(777.11)) on node run container_id:roottestasyncloaddatabases-gw7-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw7-node1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query INSERT INTO blob_storage_table VALUES (0,'data'),(1,'data') on node1 Stdout:1581 Executing query select dictGet('LocalClickHouse_direct_', 'UInt64_', toUInt64(2)) on local_node Executing query DROP TABLE IF EXISTS tbl ON CLUSTER 'cluster3' SYNC on node1 [gw2] PASSED test_backup_restore_on_cluster/test.py::test_table_with_parts_in_queue_considered_non_empty Executing query select dictGetFloat64OrDefault('MySQL_direct_', 'Float64_', toUInt64(9999999), toFloat64(777.11)) on node Executing query SYSTEM SYNC REPLICA blob_storage_table on node2 Executing query select dictGetUInt64('LocalClickHouse_direct_', 'UInt64_', toUInt64(2)) on local_node Executing query DROP TABLE IF EXISTS tbl2 ON CLUSTER 'cluster3' SYNC on node1 Executing query select dictGet('MySQL_direct_', 'ParentKeyField', toUInt64(1)) on node Executing query SELECT * FROM blob_storage_table order by id FORMAT Values on node1 Executing query select dictHas('LocalClickHouse_direct_', toUInt64(2)) on local_node Executing query select dictGetUInt64('MySQL_direct_', 'ParentKeyField', toUInt64(1)) on node Executing query SELECT * FROM blob_storage_table order by id FORMAT Values on node2 Executing query DROP DATABASE IF EXISTS mydb ON CLUSTER 'cluster3' SYNC on node1 run container_id:roottestasyncloaddatabases-gw7-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw7-node1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:1581 Executing query select dictGetOrDefault('LocalClickHouse_direct_', 'UInt64_', toUInt64(9999999), toUInt64(88)) on local_node Executing query select dictHas('MySQL_direct_', toUInt64(1)) on node Request URL: 'http://127.0.0.1:30400/devstoreaccount1/cont?restype=REDACTED&comp=REDACTED' Request method: 'GET' Request headers: 'x-ms-version': 'REDACTED' 'Accept': 'application/xml' 'User-Agent': 'azsdk-python-storage-blob/12.19.0 Python/3.10.12 (Linux-5.15.0-130-generic-x86_64-with-glibc2.35)' 'x-ms-date': 'REDACTED' 'x-ms-client-request-id': '3d59e820-0f69-11f0-bbd2-0242ac110002' 'Authorization': 'REDACTED' No body was attached to the request http://127.0.0.1:30400 "GET /devstoreaccount1/cont?restype=container&comp=list HTTP/1.1" 200 None Response status: 200 Response headers: 'Server': 'Azurite-Blob/3.34.0' 'content-type': 'application/xml' 'x-ms-client-request-id': '3d59e820-0f69-11f0-bbd2-0242ac110002' 'x-ms-request-id': '520c8788-11a7-48a2-88b3-c1ae88a07d45' 'x-ms-version': 'REDACTED' 'date': 'Wed, 02 Apr 2025 02:21:52 GMT' 'Connection': 'keep-alive' 'Keep-Alive': 'REDACTED' 'Transfer-Encoding': 'chunked' Executing query INSERT INTO blob_storage_table VALUES (2,'data'),(3,'data') on node2 Executing query DROP DATABASE IF EXISTS mydb2 ON CLUSTER 'cluster3' SYNC on node1 Executing query select dictGetUInt64OrDefault('LocalClickHouse_direct_', 'UInt64_', toUInt64(9999999), toUInt64(88)) on local_node Executing query select dictGetOrDefault('MySQL_direct_', 'ParentKeyField', toUInt64(9999999), toUInt64(444)) on node Executing query SYSTEM SYNC REPLICA blob_storage_table on node1 Executing query select dictGetUInt64OrDefault('MySQL_direct_', 'ParentKeyField', toUInt64(9999999), toUInt64(444)) on node Executing query DROP USER IF EXISTS u1, u2 ON CLUSTER 'cluster3' on node1 Executing query select dictGet('LocalClickHouse_direct_', 'Int8_', toUInt64(2)) on local_node Executing query SELECT * FROM blob_storage_table order by id FORMAT Values on node2 Executing query select dictGet('MySQL_direct_', 'UInt8_', toUInt64(2)) on node Executing query SELECT * FROM blob_storage_table order by id FORMAT Values on node1 Executing query select dictGetInt8('LocalClickHouse_direct_', 'Int8_', toUInt64(2)) on local_node test_backup_restore_on_cluster/test.py::test_tables_dependency Executing query CREATE DATABASE mydb ON CLUSTER 'cluster3' on node1 run container_id:roottestasyncloaddatabases-gw7-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw7-node1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:1581 Executing query select dictGetUInt8('MySQL_direct_', 'UInt8_', toUInt64(2)) on node Request URL: 'http://127.0.0.1:30400/devstoreaccount1/cont?restype=REDACTED&comp=REDACTED' Request method: 'GET' Request headers: 'x-ms-version': 'REDACTED' 'Accept': 'application/xml' 'User-Agent': 'azsdk-python-storage-blob/12.19.0 Python/3.10.12 (Linux-5.15.0-130-generic-x86_64-with-glibc2.35)' 'x-ms-date': 'REDACTED' 'x-ms-client-request-id': '3dfe5518-0f69-11f0-bbd2-0242ac110002' 'Authorization': 'REDACTED' No body was attached to the request http://127.0.0.1:30400 "GET /devstoreaccount1/cont?restype=container&comp=list HTTP/1.1" 200 None Response status: 200 Response headers: 'Server': 'Azurite-Blob/3.34.0' 'content-type': 'application/xml' 'x-ms-client-request-id': '3dfe5518-0f69-11f0-bbd2-0242ac110002' 'x-ms-request-id': '0f14cc12-a1ce-42f0-b785-b727ac1d3f3e' 'x-ms-version': 'REDACTED' 'date': 'Wed, 02 Apr 2025 02:21:53 GMT' 'Connection': 'keep-alive' 'Keep-Alive': 'REDACTED' 'Transfer-Encoding': 'chunked' Executing query DROP TABLE blob_storage_table ON CLUSTER test_cluster SYNC on node1 Executing query select dictHas('LocalClickHouse_direct_', toUInt64(2)) on local_node Executing query CREATE TABLE mydb.src ON CLUSTER 'cluster' (x Int64, y String) ENGINE=MergeTree ORDER BY tuple() on node1 Executing query select dictHas('MySQL_direct_', toUInt64(2)) on node Executing query select dictGetOrDefault('LocalClickHouse_direct_', 'Int8_', toUInt64(9999999), toInt8(-55)) on local_node Command:[docker compose --env-file /ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw8/.env --project-name roottestazureblobstoragezerocopyreplication-gw8 --file /ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw8/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_azurite.yml --file /ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw8/node2/docker-compose.yml stop --timeout 20] [gw8] PASSED test_azure_blob_storage_zero_copy_replication/test.py::test_zero_copy_replication Executing query select dictGetOrDefault('MySQL_direct_', 'UInt8_', toUInt64(9999999), toUInt8(55)) on node Executing query CREATE DICTIONARY mydb.dict ON CLUSTER 'cluster' (x Int64, y String) PRIMARY KEY x SOURCE(CLICKHOUSE(HOST 'localhost' PORT tcpPort() DB 'mydb' TABLE 'src')) LAYOUT(FLAT()) LIFETIME(0) on node1 Executing query select dictGetInt8OrDefault('LocalClickHouse_direct_', 'Int8_', toUInt64(9999999), toInt8(-55)) on local_node Executing query select dictGetUInt8OrDefault('MySQL_direct_', 'UInt8_', toUInt64(9999999), toUInt8(55)) on node run container_id:roottestasyncloaddatabases-gw7-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw7-node1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] run container_id:roottestasyncloaddatabases-gw7-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw7-node1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query CREATE TABLE mydb.dist1 (x Int64) ENGINE=Distributed('cluster', 'mydb', 'src') on node1 No clickhouse process running. Start new one. http://localhost:None "POST /v1.46/containers/roottestasyncloaddatabases-gw7-node1-1/exec HTTP/1.1" 201 74 Executing query select dictGet('LocalClickHouse_direct_', 'Int16_', toUInt64(2)) on local_node http://localhost:None "POST /v1.46/exec/c6ffcd4c3a7da9ca9bff8afeaec216d1b5c85b59c6abdc72206cd63e2e56feb9/start HTTP/1.1" 200 0 http://localhost:None "GET /v1.46/exec/c6ffcd4c3a7da9ca9bff8afeaec216d1b5c85b59c6abdc72206cd63e2e56feb9/json HTTP/1.1" 200 586 Executing query select dictGet('MySQL_direct_', 'UInt16_', toUInt64(2)) on node Executing query CREATE TABLE mydb.dist2 (x Int64) ENGINE=Distributed(cluster, 'mydb', 'src') on node3 Executing query select dictGetInt16('LocalClickHouse_direct_', 'Int16_', toUInt64(2)) on local_node Executing query CREATE TABLE mydb.clusterfunc1 AS cluster('cluster', 'mydb.src') on node1 Executing query select dictGetUInt16('MySQL_direct_', 'UInt16_', toUInt64(2)) on node Executing query select dictHas('LocalClickHouse_direct_', toUInt64(2)) on local_node Executing query CREATE TABLE mydb.clusterfunc2 AS cluster(cluster, mydb.src) on node1 Executing query select dictHas('MySQL_direct_', toUInt64(2)) on node Executing query CREATE TABLE mydb.clusterfunc3 AS cluster(cluster, 'mydb', 'src') on node1 Executing query select dictGetOrDefault('LocalClickHouse_direct_', 'Int16_', toUInt64(9999999), toInt16(-66)) on local_node run container_id:roottestasyncloaddatabases-gw7-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw7-node1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Stdout:2385 Clickhouse process running. run container_id:roottestasyncloaddatabases-gw7-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] Command:[docker exec roottestasyncloaddatabases-gw7-node1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] Executing query select dictGetOrDefault('MySQL_direct_', 'UInt16_', toUInt64(9999999), toUInt16(66)) on node Stdout:2385 Executing query select 20 on node1 Executing query CREATE TABLE mydb.clusterfunc4 AS cluster(cluster, dictionary(mydb.dict)) on node1 Executing query select dictGetInt16OrDefault('LocalClickHouse_direct_', 'Int16_', toUInt64(9999999), toInt16(-66)) on local_node Executing query select dictGetUInt16OrDefault('MySQL_direct_', 'UInt16_', toUInt64(9999999), toUInt16(66)) on node Executing query CREATE TABLE mydb.clusterfunc5 AS clusterAllReplicas(cluster, dictionary(mydb.dict)) on node1 Executing query select dictGet('LocalClickHouse_direct_', 'Int32_', toUInt64(2)) on local_node Executing query select dictGet('MySQL_direct_', 'UInt32_', toUInt64(2)) on node Executing query CREATE TABLE mydb.clusterfunc6 AS cluster('cluster', 'mydb.src') on node3 Executing query select 20 on node1 Executing query CREATE TABLE mydb.clusterfunc7 AS cluster(cluster, mydb.src) on node3 Executing query select dictGetUInt32('MySQL_direct_', 'UInt32_', toUInt64(2)) on node Executing query select dictGetInt32('LocalClickHouse_direct_', 'Int32_', toUInt64(2)) on local_node Executing query CREATE TABLE mydb.clusterfunc8 AS cluster(cluster, 'mydb', 'src') on node3 Executing query select dictHas('MySQL_direct_', toUInt64(2)) on node Executing query select dictHas('LocalClickHouse_direct_', toUInt64(2)) on local_node Executing query CREATE TABLE mydb.clusterfunc9 AS cluster(cluster, dictionary(mydb.dict)) on node3 Executing query select 20 on node1 Executing query CREATE TABLE mydb.clusterfunc10 AS clusterAllReplicas(cluster, dictionary(mydb.dict)) on node3 Executing query select dictGetOrDefault('MySQL_direct_', 'UInt32_', toUInt64(9999999), toUInt32(77)) on node Executing query select dictGetOrDefault('LocalClickHouse_direct_', 'Int32_', toUInt64(9999999), toInt32(-77)) on local_node Executing query BACKUP DATABASE mydb ON CLUSTER 'cluster3' TO Disk('backups', '31') on node3 Executing query select dictGetUInt32OrDefault('MySQL_direct_', 'UInt32_', toUInt64(9999999), toUInt32(77)) on node Executing query select dictGetInt32OrDefault('LocalClickHouse_direct_', 'Int32_', toUInt64(9999999), toInt32(-77)) on local_node Executing query select dictGet('MySQL_direct_', 'UInt64_', toUInt64(2)) on node Executing query select 20 on node1 Executing query select dictGet('LocalClickHouse_direct_', 'Int64_', toUInt64(2)) on local_node Executing query select dictGetUInt64('MySQL_direct_', 'UInt64_', toUInt64(2)) on node Executing query select dictGetInt64('LocalClickHouse_direct_', 'Int64_', toUInt64(2)) on local_node Executing query DROP DATABASE mydb on node3 Executing query select dictHas('MySQL_direct_', toUInt64(2)) on node Executing query RESTORE DATABASE mydb ON CLUSTER 'cluster3' FROM Disk('backups', '31') on node3 Executing query select dictHas('LocalClickHouse_direct_', toUInt64(2)) on local_node Executing query select 20 on node1 Executing query select dictGetOrDefault('MySQL_direct_', 'UInt64_', toUInt64(9999999), toUInt64(88)) on node Executing query select dictGetOrDefault('LocalClickHouse_direct_', 'Int64_', toUInt64(9999999), toInt64(-88)) on local_node Executing query select dictGetUInt64OrDefault('MySQL_direct_', 'UInt64_', toUInt64(9999999), toUInt64(88)) on node Executing query SYSTEM FLUSH LOGS ON CLUSTER 'cluster3' on node3 Executing query select dictGetInt64OrDefault('LocalClickHouse_direct_', 'Int64_', toUInt64(9999999), toInt64(-88)) on local_node Executing query select 20 on node1 Executing query select dictGet('MySQL_direct_', 'Int8_', toUInt64(2)) on node Executing query select dictGet('LocalClickHouse_direct_', 'UUID_', toUInt64(2)) on local_node Executing query select count() from test.table_8 on node1 Executing query select dictGetInt8('MySQL_direct_', 'Int8_', toUInt64(2)) on node Executing query select dictGetUUID('LocalClickHouse_direct_', 'UUID_', toUInt64(2)) on local_node Executing query select count() from test.table_9 on node1 Executing query select dictHas('MySQL_direct_', toUInt64(2)) on node Executing query select count() from test.table_6 on node1 Executing query select dictHas('LocalClickHouse_direct_', toUInt64(2)) on local_node Executing query select dictGetOrDefault('MySQL_direct_', 'Int8_', toUInt64(9999999), toInt8(-55)) on node Executing query select count() from test.table_2 on node1 Executing query select dictGetOrDefault('LocalClickHouse_direct_', 'UUID_', toUInt64(9999999), toUUID('550e8400-0000-0000-0000-000000000000')) on local_node run container_id:roottestbackuprestoreoncluster-gw2-node1-1 detach:False nothrow:False cmd: ['bash', '-c', '[ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "RestorerFromBackup: Table mydb.src has no dependencies (level 0)" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "" ] && cat || grep -v "$" ) || true'] Command:[docker exec roottestbackuprestoreoncluster-gw2-node1-1 bash -c [ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "RestorerFromBackup: Table mydb.src has no dependencies (level 0)" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "" ] && cat || grep -v "$" ) || true] Executing query select dictGetInt8OrDefault('MySQL_direct_', 'Int8_', toUInt64(9999999), toInt8(-55)) on node Stdout:/var/log/clickhouse-server/clickhouse-server.log:2025.04.02 02:21:58.513173 [ 721 ] {2d1af6d6-2e5a-483f-8ee1-cfae83911ffa} RestorerFromBackup: Table mydb.src has no dependencies (level 0) run container_id:roottestbackuprestoreoncluster-gw2-node1-1 detach:False nothrow:False cmd: ['bash', '-c', '[ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "RestorerFromBackup: Table mydb.dict has 1 dependencies: mydb.src (level 1)" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "" ] && cat || grep -v "$" ) || true'] Command:[docker exec roottestbackuprestoreoncluster-gw2-node1-1 bash -c [ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "RestorerFromBackup: Table mydb.dict has 1 dependencies: mydb.src (level 1)" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "" ] && cat || grep -v "$" ) || true] Stdout:/var/log/clickhouse-server/clickhouse-server.log:2025.04.02 02:21:58.513318 [ 721 ] {2d1af6d6-2e5a-483f-8ee1-cfae83911ffa} RestorerFromBackup: Table mydb.dict has 1 dependencies: mydb.src (level 1) run container_id:roottestbackuprestoreoncluster-gw2-node1-1 detach:False nothrow:False cmd: ['bash', '-c', '[ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "RestorerFromBackup: Table mydb.dist1 has 1 dependencies: mydb.src (level 1)" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "" ] && cat || grep -v "$" ) || true'] Command:[docker exec roottestbackuprestoreoncluster-gw2-node1-1 bash -c [ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "RestorerFromBackup: Table mydb.dist1 has 1 dependencies: mydb.src (level 1)" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "" ] && cat || grep -v "$" ) || true] Executing query select count() from test.table_19 on node1 Stdout:/var/log/clickhouse-server/clickhouse-server.log:2025.04.02 02:21:58.513283 [ 721 ] {2d1af6d6-2e5a-483f-8ee1-cfae83911ffa} RestorerFromBackup: Table mydb.dist1 has 1 dependencies: mydb.src (level 1) run container_id:roottestbackuprestoreoncluster-gw2-node1-1 detach:False nothrow:False cmd: ['bash', '-c', '[ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "RestorerFromBackup: Table mydb.clusterfunc1 has 1 dependencies: mydb.src (level 1)" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "" ] && cat || grep -v "$" ) || true'] Command:[docker exec roottestbackuprestoreoncluster-gw2-node1-1 bash -c [ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "RestorerFromBackup: Table mydb.clusterfunc1 has 1 dependencies: mydb.src (level 1)" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "" ] && cat || grep -v "$" ) || true] Executing query select dictGetUUIDOrDefault('LocalClickHouse_direct_', 'UUID_', toUInt64(9999999), toUUID('550e8400-0000-0000-0000-000000000000')) on local_node Executing query select dictGet('MySQL_direct_', 'Int16_', toUInt64(2)) on node Stdout:/var/log/clickhouse-server/clickhouse-server.log:2025.04.02 02:21:58.513231 [ 721 ] {2d1af6d6-2e5a-483f-8ee1-cfae83911ffa} RestorerFromBackup: Table mydb.clusterfunc1 has 1 dependencies: mydb.src (level 1) run container_id:roottestbackuprestoreoncluster-gw2-node1-1 detach:False nothrow:False cmd: ['bash', '-c', '[ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "RestorerFromBackup: Table mydb.clusterfunc2 has 1 dependencies: mydb.src (level 1)" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "" ] && cat || grep -v "$" ) || true'] Command:[docker exec roottestbackuprestoreoncluster-gw2-node1-1 bash -c [ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "RestorerFromBackup: Table mydb.clusterfunc2 has 1 dependencies: mydb.src (level 1)" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "" ] && cat || grep -v "$" ) || true] Stdout:/var/log/clickhouse-server/clickhouse-server.log:2025.04.02 02:21:58.513402 [ 721 ] {2d1af6d6-2e5a-483f-8ee1-cfae83911ffa} RestorerFromBackup: Table mydb.clusterfunc2 has 1 dependencies: mydb.src (level 1) run container_id:roottestbackuprestoreoncluster-gw2-node1-1 detach:False nothrow:False cmd: ['bash', '-c', '[ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "RestorerFromBackup: Table mydb.clusterfunc3 has 1 dependencies: mydb.src (level 1)" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "" ] && cat || grep -v "$" ) || true'] Command:[docker exec roottestbackuprestoreoncluster-gw2-node1-1 bash -c [ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "RestorerFromBackup: Table mydb.clusterfunc3 has 1 dependencies: mydb.src (level 1)" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "" ] && cat || grep -v "$" ) || true] Stdout:/var/log/clickhouse-server/clickhouse-server.log:2025.04.02 02:21:58.513354 [ 721 ] {2d1af6d6-2e5a-483f-8ee1-cfae83911ffa} RestorerFromBackup: Table mydb.clusterfunc3 has 1 dependencies: mydb.src (level 1) run container_id:roottestbackuprestoreoncluster-gw2-node1-1 detach:False nothrow:False cmd: ['bash', '-c', '[ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "RestorerFromBackup: Table mydb.clusterfunc4 has 1 dependencies: mydb.dict (level 2)" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "" ] && cat || grep -v "$" ) || true'] Command:[docker exec roottestbackuprestoreoncluster-gw2-node1-1 bash -c [ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "RestorerFromBackup: Table mydb.clusterfunc4 has 1 dependencies: mydb.dict (level 2)" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "" ] && cat || grep -v "$" ) || true] Executing query select count() from test.table_15 on node1 Executing query select dictGetInt16('MySQL_direct_', 'Int16_', toUInt64(2)) on node Stdout:/var/log/clickhouse-server/clickhouse-server.log:2025.04.02 02:21:58.513470 [ 721 ] {2d1af6d6-2e5a-483f-8ee1-cfae83911ffa} RestorerFromBackup: Table mydb.clusterfunc4 has 1 dependencies: mydb.dict (level 2) run container_id:roottestbackuprestoreoncluster-gw2-node1-1 detach:False nothrow:False cmd: ['bash', '-c', '[ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "RestorerFromBackup: Table mydb.clusterfunc5 has 1 dependencies: mydb.dict (level 2)" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "" ] && cat || grep -v "$" ) || true'] Command:[docker exec roottestbackuprestoreoncluster-gw2-node1-1 bash -c [ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "RestorerFromBackup: Table mydb.clusterfunc5 has 1 dependencies: mydb.dict (level 2)" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "" ] && cat || grep -v "$" ) || true] Executing query select dictGet('LocalClickHouse_direct_', 'Date_', toUInt64(2)) on local_node Stdout:/var/log/clickhouse-server/clickhouse-server.log:2025.04.02 02:21:58.513436 [ 721 ] {2d1af6d6-2e5a-483f-8ee1-cfae83911ffa} RestorerFromBackup: Table mydb.clusterfunc5 has 1 dependencies: mydb.dict (level 2) run container_id:roottestbackuprestoreoncluster-gw2-node2-1 detach:False nothrow:False cmd: ['bash', '-c', '[ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "RestorerFromBackup: Table mydb.src has no dependencies (level 0)" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "" ] && cat || grep -v "$" ) || true'] Command:[docker exec roottestbackuprestoreoncluster-gw2-node2-1 bash -c [ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "RestorerFromBackup: Table mydb.src has no dependencies (level 0)" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "" ] && cat || grep -v "$" ) || true] Stdout:/var/log/clickhouse-server/clickhouse-server.log:2025.04.02 02:21:58.510273 [ 1568 ] {431ce058-bcef-4b46-a8c2-576d362d9f0c} RestorerFromBackup: Table mydb.src has no dependencies (level 0) run container_id:roottestbackuprestoreoncluster-gw2-node2-1 detach:False nothrow:False cmd: ['bash', '-c', '[ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "RestorerFromBackup: Table mydb.dict has 1 dependencies: mydb.src (level 1)" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "" ] && cat || grep -v "$" ) || true'] Command:[docker exec roottestbackuprestoreoncluster-gw2-node2-1 bash -c [ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "RestorerFromBackup: Table mydb.dict has 1 dependencies: mydb.src (level 1)" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "" ] && cat || grep -v "$" ) || true] Stdout:/var/log/clickhouse-server/clickhouse-server.log:2025.04.02 02:21:58.510331 [ 1568 ] {431ce058-bcef-4b46-a8c2-576d362d9f0c} RestorerFromBackup: Table mydb.dict has 1 dependencies: mydb.src (level 1) run container_id:roottestbackuprestoreoncluster-gw2-node3-1 detach:False nothrow:False cmd: ['bash', '-c', '[ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "RestorerFromBackup: Table mydb.dist2 has no dependencies (level 0)" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "" ] && cat || grep -v "$" ) || true'] Command:[docker exec roottestbackuprestoreoncluster-gw2-node3-1 bash -c [ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "RestorerFromBackup: Table mydb.dist2 has no dependencies (level 0)" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "" ] && cat || grep -v "$" ) || true] Executing query select count() from test.table_10 on node1 Stdout:/var/log/clickhouse-server/clickhouse-server.log:2025.04.02 02:21:58.514031 [ 689 ] {60885311-a9e0-4209-97c4-817f41df4dc5} RestorerFromBackup: Table mydb.dist2 has no dependencies (level 0) run container_id:roottestbackuprestoreoncluster-gw2-node3-1 detach:False nothrow:False cmd: ['bash', '-c', '[ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "RestorerFromBackup: Table mydb.clusterfunc6 has no dependencies (level 0)" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "" ] && cat || grep -v "$" ) || true'] Command:[docker exec roottestbackuprestoreoncluster-gw2-node3-1 bash -c [ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "RestorerFromBackup: Table mydb.clusterfunc6 has no dependencies (level 0)" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "" ] && cat || grep -v "$" ) || true] Stdout:/var/log/clickhouse-server/clickhouse-server.log:2025.04.02 02:21:58.514209 [ 689 ] {60885311-a9e0-4209-97c4-817f41df4dc5} RestorerFromBackup: Table mydb.clusterfunc6 has no dependencies (level 0) run container_id:roottestbackuprestoreoncluster-gw2-node3-1 detach:False nothrow:False cmd: ['bash', '-c', '[ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "RestorerFromBackup: Table mydb.clusterfunc7 has no dependencies (level 0)" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "" ] && cat || grep -v "$" ) || true'] Command:[docker exec roottestbackuprestoreoncluster-gw2-node3-1 bash -c [ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "RestorerFromBackup: Table mydb.clusterfunc7 has no dependencies (level 0)" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "" ] && cat || grep -v "$" ) || true] Executing query select dictGetDate('LocalClickHouse_direct_', 'Date_', toUInt64(2)) on local_node Executing query select dictHas('MySQL_direct_', toUInt64(2)) on node Stdout:/var/log/clickhouse-server/clickhouse-server.log:2025.04.02 02:21:58.514179 [ 689 ] {60885311-a9e0-4209-97c4-817f41df4dc5} RestorerFromBackup: Table mydb.clusterfunc7 has no dependencies (level 0) run container_id:roottestbackuprestoreoncluster-gw2-node3-1 detach:False nothrow:False cmd: ['bash', '-c', '[ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "RestorerFromBackup: Table mydb.clusterfunc8 has no dependencies (level 0)" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "" ] && cat || grep -v "$" ) || true'] Command:[docker exec roottestbackuprestoreoncluster-gw2-node3-1 bash -c [ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "RestorerFromBackup: Table mydb.clusterfunc8 has no dependencies (level 0)" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "" ] && cat || grep -v "$" ) || true] Stdout:/var/log/clickhouse-server/clickhouse-server.log:2025.04.02 02:21:58.514076 [ 689 ] {60885311-a9e0-4209-97c4-817f41df4dc5} RestorerFromBackup: Table mydb.clusterfunc8 has no dependencies (level 0) run container_id:roottestbackuprestoreoncluster-gw2-node3-1 detach:False nothrow:False cmd: ['bash', '-c', '[ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "RestorerFromBackup: Table mydb.clusterfunc9 has no dependencies (level 0)" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "" ] && cat || grep -v "$" ) || true'] Command:[docker exec roottestbackuprestoreoncluster-gw2-node3-1 bash -c [ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "RestorerFromBackup: Table mydb.clusterfunc9 has no dependencies (level 0)" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "" ] && cat || grep -v "$" ) || true] Executing query select count() from test.table_18 on node1 Stdout:/var/log/clickhouse-server/clickhouse-server.log:2025.04.02 02:21:58.513965 [ 689 ] {60885311-a9e0-4209-97c4-817f41df4dc5} RestorerFromBackup: Table mydb.clusterfunc9 has no dependencies (level 0) run container_id:roottestbackuprestoreoncluster-gw2-node3-1 detach:False nothrow:False cmd: ['bash', '-c', '[ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "RestorerFromBackup: Table mydb.clusterfunc10 has no dependencies (level 0)" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "" ] && cat || grep -v "$" ) || true'] Command:[docker exec roottestbackuprestoreoncluster-gw2-node3-1 bash -c [ -f /var/log/clickhouse-server/clickhouse-server.log ] && zgrep -aH "RestorerFromBackup: Table mydb.clusterfunc10 has no dependencies (level 0)" /var/log/clickhouse-server/clickhouse-server.log | ( [ -z "" ] && cat || grep -v "$" ) || true] Stdout:/var/log/clickhouse-server/clickhouse-server.log:2025.04.02 02:21:58.513904 [ 689 ] {60885311-a9e0-4209-97c4-817f41df4dc5} RestorerFromBackup: Table mydb.clusterfunc10 has no dependencies (level 0) Executing query DROP TABLE IF EXISTS tbl ON CLUSTER 'cluster3' SYNC on node1 [gw2] PASSED test_backup_restore_on_cluster/test.py::test_tables_dependency Executing query select dictHas('LocalClickHouse_direct_', toUInt64(2)) on local_node Executing query select dictGetOrDefault('MySQL_direct_', 'Int16_', toUInt64(9999999), toInt16(-66)) on node Executing query select count() from test.table_0 on node1 Executing query select dictGetInt16OrDefault('MySQL_direct_', 'Int16_', toUInt64(9999999), toInt16(-66)) on node Executing query DROP TABLE IF EXISTS tbl2 ON CLUSTER 'cluster3' SYNC on node1 Executing query select dictGetOrDefault('LocalClickHouse_direct_', 'Date_', toUInt64(9999999), toDate('2018-12-30')) on local_node Executing query select count() from test.table_17 on node1 Executing query select dictGet('MySQL_direct_', 'Int32_', toUInt64(2)) on node Executing query select dictGetDateOrDefault('LocalClickHouse_direct_', 'Date_', toUInt64(9999999), toDate('2018-12-30')) on local_node Executing query DROP DATABASE IF EXISTS mydb ON CLUSTER 'cluster3' SYNC on node1 Executing query select count() from test.table_7 on node1 Executing query select dictGetInt32('MySQL_direct_', 'Int32_', toUInt64(2)) on node Executing query select dictGet('LocalClickHouse_direct_', 'DateTime_', toUInt64(2)) on local_node Executing query DROP DATABASE IF EXISTS mydb2 ON CLUSTER 'cluster3' SYNC on node1 Executing query select dictHas('MySQL_direct_', toUInt64(2)) on node Executing query select count() from test.table_14 on node1 Executing query select dictGetDateTime('LocalClickHouse_direct_', 'DateTime_', toUInt64(2)) on local_node Executing query select count() from test.table_11 on node1 Executing query select dictGetOrDefault('MySQL_direct_', 'Int32_', toUInt64(9999999), toInt32(-77)) on node Executing query select dictHas('LocalClickHouse_direct_', toUInt64(2)) on local_node Executing query DROP USER IF EXISTS u1, u2 ON CLUSTER 'cluster3' on node1 Executing query select count() from test.table_13 on node1 Executing query select dictGetInt32OrDefault('MySQL_direct_', 'Int32_', toUInt64(9999999), toInt32(-77)) on node Executing query select dictGetOrDefault('LocalClickHouse_direct_', 'DateTime_', toUInt64(9999999), toDateTime('2018-12-30 00:00:00')) on local_node Command:[docker compose --env-file /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw2/.env --project-name roottestbackuprestoreoncluster-gw2 --file /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw2/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw2/node2/docker-compose.yml --file /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw2/node3/docker-compose.yml stop --timeout 20] Executing query select count() from test.table_12 on node1 Executing query select dictGet('MySQL_direct_', 'Int64_', toUInt64(2)) on node Executing query select dictGetDateTimeOrDefault('LocalClickHouse_direct_', 'DateTime_', toUInt64(9999999), toDateTime('2018-12-30 00:00:00')) on local_node Executing query select count() from test.table_3 on node1 Executing query select dictGetInt64('MySQL_direct_', 'Int64_', toUInt64(2)) on node Executing query select dictGet('LocalClickHouse_direct_', 'String_', toUInt64(2)) on local_node Executing query select count() from test.table_4 on node1 Executing query select dictHas('MySQL_direct_', toUInt64(2)) on node Executing query select dictGetString('LocalClickHouse_direct_', 'String_', toUInt64(2)) on local_node Executing query select dictGetOrDefault('MySQL_direct_', 'Int64_', toUInt64(9999999), toInt64(-88)) on node Executing query select count() from test.table_16 on node1 Executing query select dictHas('LocalClickHouse_direct_', toUInt64(2)) on local_node Executing query select count() from test.table_5 on node1 Executing query select dictGetInt64OrDefault('MySQL_direct_', 'Int64_', toUInt64(9999999), toInt64(-88)) on node Executing query select dictGetOrDefault('LocalClickHouse_direct_', 'String_', toUInt64(9999999), toString('hi')) on local_node Executing query select count() from test.table_1 on node1 Executing query select dictGet('MySQL_direct_', 'UUID_', toUInt64(2)) on node Executing query select dictGetStringOrDefault('LocalClickHouse_direct_', 'String_', toUInt64(9999999), toString('hi')) on local_node Executing query select dictGetUUID('MySQL_direct_', 'UUID_', toUInt64(2)) on node Executing query drop table test.table_0 sync on node1 Executing query select dictGet('LocalClickHouse_direct_', 'Float32_', toUInt64(2)) on local_node Executing query drop table test.table_1 sync on node1 Executing query select dictHas('MySQL_direct_', toUInt64(2)) on node Executing query drop table test.table_2 sync on node1 Executing query select dictGetFloat32('LocalClickHouse_direct_', 'Float32_', toUInt64(2)) on local_node Executing query select dictGetOrDefault('MySQL_direct_', 'UUID_', toUInt64(9999999), toUUID('550e8400-0000-0000-0000-000000000000')) on node Executing query drop table test.table_3 sync on node1 Executing query select dictHas('LocalClickHouse_direct_', toUInt64(2)) on local_node Executing query select dictGetUUIDOrDefault('MySQL_direct_', 'UUID_', toUInt64(9999999), toUUID('550e8400-0000-0000-0000-000000000000')) on node Executing query drop table test.table_4 sync on node1 Executing query drop table test.table_5 sync on node1 Executing query select dictGetOrDefault('LocalClickHouse_direct_', 'Float32_', toUInt64(9999999), toFloat32(555.11)) on local_node Executing query select dictGet('MySQL_direct_', 'Date_', toUInt64(2)) on node Executing query drop table test.table_6 sync on node1 Executing query select dictGetFloat32OrDefault('LocalClickHouse_direct_', 'Float32_', toUInt64(9999999), toFloat32(555.11)) on local_node Executing query select dictGetDate('MySQL_direct_', 'Date_', toUInt64(2)) on node Executing query drop table test.table_7 sync on node1 Executing query select dictHas('MySQL_direct_', toUInt64(2)) on node Executing query select dictGet('LocalClickHouse_direct_', 'Float64_', toUInt64(2)) on local_node Executing query drop table test.table_8 sync on node1 Executing query select dictGetOrDefault('MySQL_direct_', 'Date_', toUInt64(9999999), toDate('2018-12-30')) on node Executing query drop table test.table_9 sync on node1 Executing query select dictGetFloat64('LocalClickHouse_direct_', 'Float64_', toUInt64(2)) on local_node Executing query drop table test.table_10 sync on node1 Executing query select dictGetDateOrDefault('MySQL_direct_', 'Date_', toUInt64(9999999), toDate('2018-12-30')) on node Executing query select dictHas('LocalClickHouse_direct_', toUInt64(2)) on local_node Executing query drop table test.table_11 sync on node1 Executing query select dictGet('MySQL_direct_', 'DateTime_', toUInt64(2)) on node Executing query select dictGetOrDefault('LocalClickHouse_direct_', 'Float64_', toUInt64(9999999), toFloat64(777.11)) on local_node Executing query drop table test.table_12 sync on node1 Executing query select dictGetDateTime('MySQL_direct_', 'DateTime_', toUInt64(2)) on node Executing query select dictGetFloat64OrDefault('LocalClickHouse_direct_', 'Float64_', toUInt64(9999999), toFloat64(777.11)) on local_node Executing query drop table test.table_13 sync on node1 Executing query select dictHas('MySQL_direct_', toUInt64(2)) on node Executing query drop table test.table_14 sync on node1 Executing query select dictGet('LocalClickHouse_direct_', 'ParentKeyField', toUInt64(2)) on local_node Executing query select dictGetOrDefault('MySQL_direct_', 'DateTime_', toUInt64(9999999), toDateTime('2018-12-30 00:00:00')) on node Executing query drop table test.table_15 sync on node1 Executing query select dictGetDateTimeOrDefault('MySQL_direct_', 'DateTime_', toUInt64(9999999), toDateTime('2018-12-30 00:00:00')) on node Executing query select dictGetUInt64('LocalClickHouse_direct_', 'ParentKeyField', toUInt64(2)) on local_node Stderr: Container roottestbackuprestoreoncluster-gw2-node1-1 Stopping Stderr: Container roottestbackuprestoreoncluster-gw2-node2-1 Stopping Stderr: Container roottestbackuprestoreoncluster-gw2-node3-1 Stopping Stderr: Container roottestbackuprestoreoncluster-gw2-node2-1 Stopped Stderr: Container roottestbackuprestoreoncluster-gw2-node3-1 Stopped Stderr: Container roottestbackuprestoreoncluster-gw2-node1-1 Stopped Stderr: Container roottestbackuprestoreoncluster-gw2-zoo3-1 Stopping Stderr: Container roottestbackuprestoreoncluster-gw2-zoo1-1 Stopping Stderr: Container roottestbackuprestoreoncluster-gw2-zoo2-1 Stopping Stderr: Container roottestbackuprestoreoncluster-gw2-zoo3-1 Stopped Stderr: Container roottestbackuprestoreoncluster-gw2-zoo2-1 Stopped Stderr: Container roottestbackuprestoreoncluster-gw2-zoo1-1 Stopped Command:[bash -c [ -f /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw2/node1/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw2/node1/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[bash -c [ -f /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw2/node2/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw2/node2/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[bash -c [ -f /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw2/node3/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw2/node3/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[docker compose --env-file /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw2/.env --project-name roottestbackuprestoreoncluster-gw2 --file /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw2/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw2/node2/docker-compose.yml --file /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-0-gw2/node3/docker-compose.yml down --volumes] Executing query drop table test.table_16 sync on node1 Executing query select dictGet('MySQL_direct_', 'String_', toUInt64(2)) on node Executing query select dictHas('LocalClickHouse_direct_', toUInt64(2)) on local_node Executing query drop table test.table_17 sync on node1 Executing query select dictGetString('MySQL_direct_', 'String_', toUInt64(2)) on node Executing query select dictGetOrDefault('LocalClickHouse_direct_', 'ParentKeyField', toUInt64(9999999), toUInt64(444)) on local_node Stderr: Container roottestbackuprestoreoncluster-gw2-node1-1 Stopping Stderr: Container roottestbackuprestoreoncluster-gw2-node2-1 Stopping Stderr: Container roottestbackuprestoreoncluster-gw2-node3-1 Stopping Stderr: Container roottestbackuprestoreoncluster-gw2-node2-1 Stopped Stderr: Container roottestbackuprestoreoncluster-gw2-node2-1 Removing Stderr: Container roottestbackuprestoreoncluster-gw2-node3-1 Stopped Stderr: Container roottestbackuprestoreoncluster-gw2-node3-1 Removing Stderr: Container roottestbackuprestoreoncluster-gw2-node1-1 Stopped Stderr: Container roottestbackuprestoreoncluster-gw2-node1-1 Removing Stderr: Container roottestbackuprestoreoncluster-gw2-node2-1 Removed Stderr: Container roottestbackuprestoreoncluster-gw2-node3-1 Removed Stderr: Container roottestbackuprestoreoncluster-gw2-node1-1 Removed Stderr: Container roottestbackuprestoreoncluster-gw2-zoo2-1 Stopping Stderr: Container roottestbackuprestoreoncluster-gw2-zoo3-1 Stopping Stderr: Container roottestbackuprestoreoncluster-gw2-zoo1-1 Stopping Stderr: Container roottestbackuprestoreoncluster-gw2-zoo2-1 Stopped Stderr: Container roottestbackuprestoreoncluster-gw2-zoo2-1 Removing Stderr: Container roottestbackuprestoreoncluster-gw2-zoo3-1 Stopped Stderr: Container roottestbackuprestoreoncluster-gw2-zoo3-1 Removing Stderr: Container roottestbackuprestoreoncluster-gw2-zoo1-1 Stopped Stderr: Container roottestbackuprestoreoncluster-gw2-zoo1-1 Removing Stderr: Container roottestbackuprestoreoncluster-gw2-zoo2-1 Removed Stderr: Container roottestbackuprestoreoncluster-gw2-zoo1-1 Removed Stderr: Container roottestbackuprestoreoncluster-gw2-zoo3-1 Removed Stderr: Network roottestbackuprestoreoncluster-gw2_default Removing Stderr: Network roottestbackuprestoreoncluster-gw2_default Removed Cleanup called Executing query drop table test.table_18 sync on node1 Docker networks for project roottestbackuprestoreoncluster-gw2 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestbackuprestoreoncluster-gw2 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestbackuprestoreoncluster-gw2 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestbackuprestoreoncluster-gw2-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} No running containers for project: roottestbackuprestoreoncluster-gw2 Trying to prune unused networks... Trying to prune unused images... Command:[docker image prune -f] Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Stdout:4 Command:[docker volume prune -f] Stdout:Total reclaimed space: 0B Volumes pruned: 4 Executing query select dictHas('MySQL_direct_', toUInt64(2)) on node Executing query drop table test.table_19 sync on node1 Executing query select dictGetUInt64OrDefault('LocalClickHouse_direct_', 'ParentKeyField', toUInt64(9999999), toUInt64(444)) on local_node Command:[docker compose --env-file /ClickHouse/tests/integration/test_async_load_databases/_instances-0-gw7/.env --project-name roottestasyncloaddatabases-gw7 --file /ClickHouse/tests/integration/test_async_load_databases/_instances-0-gw7/node1/docker-compose.yml --file /ClickHouse/tests/integration/test_async_load_databases/_instances-0-gw7/node2/docker-compose.yml stop --timeout 20] [gw7] PASSED test_async_load_databases/test.py::test_multiple_tables Executing query select dictGetOrDefault('MySQL_direct_', 'String_', toUInt64(9999999), toString('hi')) on node Executing query select dictGetHierarchy('LocalClickHouse_direct_', toUInt64(1)) on local_node Executing query select dictGetStringOrDefault('MySQL_direct_', 'String_', toUInt64(9999999), toString('hi')) on node Stderr: Container roottestasyncloaddatabases-gw7-node2-1 Stopping Stderr: Container roottestasyncloaddatabases-gw7-node1-1 Stopping Stderr: Container roottestasyncloaddatabases-gw7-node1-1 Stopped Stderr: Container roottestasyncloaddatabases-gw7-node2-1 Stopped Command:[bash -c [ -f /ClickHouse/tests/integration/test_async_load_databases/_instances-0-gw7/node1/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_async_load_databases/_instances-0-gw7/node1/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[bash -c [ -f /ClickHouse/tests/integration/test_async_load_databases/_instances-0-gw7/node2/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_async_load_databases/_instances-0-gw7/node2/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[docker compose --env-file /ClickHouse/tests/integration/test_async_load_databases/_instances-0-gw7/.env --project-name roottestasyncloaddatabases-gw7 --file /ClickHouse/tests/integration/test_async_load_databases/_instances-0-gw7/node1/docker-compose.yml --file /ClickHouse/tests/integration/test_async_load_databases/_instances-0-gw7/node2/docker-compose.yml down --volumes] Executing query select dictGet('MySQL_direct_', 'Float32_', toUInt64(2)) on node Executing query select dictGetHierarchy('LocalClickHouse_direct_', toUInt64(2)) on local_node Executing query select dictGetFloat32('MySQL_direct_', 'Float32_', toUInt64(2)) on node Executing query select dictIsIn('LocalClickHouse_direct_', toUInt64(1), toUInt64(2)) on local_node Stderr: Container roottestasyncloaddatabases-gw7-node2-1 Stopping Stderr: Container roottestasyncloaddatabases-gw7-node1-1 Stopping Stderr: Container roottestasyncloaddatabases-gw7-node2-1 Stopped Stderr: Container roottestasyncloaddatabases-gw7-node2-1 Removing Stderr: Container roottestasyncloaddatabases-gw7-node1-1 Stopped Stderr: Container roottestasyncloaddatabases-gw7-node1-1 Removing Stderr: Container roottestasyncloaddatabases-gw7-node1-1 Removed Stderr: Container roottestasyncloaddatabases-gw7-node2-1 Removed Stderr: Network roottestasyncloaddatabases-gw7_default Removing Stderr: Network roottestasyncloaddatabases-gw7_default Removed Cleanup called Docker networks for project roottestasyncloaddatabases-gw7 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestasyncloaddatabases-gw7 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestasyncloaddatabases-gw7 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestasyncloaddatabases-gw7-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} No running containers for project: roottestasyncloaddatabases-gw7 Trying to prune unused networks... Executing query select dictHas('MySQL_direct_', toUInt64(2)) on node Trying to prune unused images... Command:[docker image prune -f] Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Stdout:4 Command:[docker volume prune -f] Executing query select dictIsIn('LocalClickHouse_direct_', toUInt64(2), toUInt64(1)) on local_node Stdout:Total reclaimed space: 0B Volumes pruned: 4 Executing query select dictGetOrDefault('MySQL_direct_', 'Float32_', toUInt64(9999999), toFloat32(555.11)) on node [gw5] PASSED test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_simple[direct] test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_simple[flat] Executing query system reload dictionaries on local_node Executing query select dictGetFloat32OrDefault('MySQL_direct_', 'Float32_', toUInt64(9999999), toFloat32(555.11)) on node Executing query select dictGet('LocalClickHouse_flat_', 'UInt8_', toUInt64(1)) on local_node Executing query select dictGet('MySQL_direct_', 'Float64_', toUInt64(2)) on node Executing query select dictGetUInt8('LocalClickHouse_flat_', 'UInt8_', toUInt64(1)) on local_node Executing query select dictGetFloat64('MySQL_direct_', 'Float64_', toUInt64(2)) on node Executing query select dictHas('LocalClickHouse_flat_', toUInt64(1)) on local_node Executing query select dictHas('MySQL_direct_', toUInt64(2)) on node Executing query select dictGetOrDefault('LocalClickHouse_flat_', 'UInt8_', toUInt64(9999999), toUInt8(55)) on local_node Executing query select dictGetOrDefault('MySQL_direct_', 'Float64_', toUInt64(9999999), toFloat64(777.11)) on node Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-node1-1 Stopping Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-node2-1 Stopping Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-node1-1 Stopped Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-node2-1 Stopped Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-zoo1-1 Stopping Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-zoo2-1 Stopping Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-azurite1-1 Stopping Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-zoo3-1 Stopping Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-zoo3-1 Stopped Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-zoo1-1 Stopped Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-zoo2-1 Stopped Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-azurite1-1 Stopped Command:[bash -c [ -f /ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw8/node1/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw8/node1/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[bash -c [ -f /ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw8/node2/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw8/node2/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[docker compose --env-file /ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw8/.env --project-name roottestazureblobstoragezerocopyreplication-gw8 --file /ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw8/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_azurite.yml --file /ClickHouse/tests/integration/test_azure_blob_storage_zero_copy_replication/_instances-0-gw8/node2/docker-compose.yml down --volumes] Executing query select dictGetUInt8OrDefault('LocalClickHouse_flat_', 'UInt8_', toUInt64(9999999), toUInt8(55)) on local_node Executing query select dictGetFloat64OrDefault('MySQL_direct_', 'Float64_', toUInt64(9999999), toFloat64(777.11)) on node Executing query select dictGet('LocalClickHouse_flat_', 'UInt16_', toUInt64(1)) on local_node Executing query select dictGet('MySQL_direct_', 'ParentKeyField', toUInt64(2)) on node Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-node1-1 Stopping Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-node2-1 Stopping Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-node1-1 Stopped Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-node1-1 Removing Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-node2-1 Stopped Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-node2-1 Removing Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-node1-1 Removed Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-node2-1 Removed Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-zoo3-1 Stopping Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-azurite1-1 Stopping Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-zoo2-1 Stopping Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-zoo1-1 Stopping Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-azurite1-1 Stopped Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-azurite1-1 Removing Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-zoo3-1 Stopped Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-zoo3-1 Removing Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-zoo1-1 Stopped Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-zoo1-1 Removing Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-zoo2-1 Stopped Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-zoo2-1 Removing Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-zoo1-1 Removed Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-zoo3-1 Removed Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-azurite1-1 Removed Stderr: Container roottestazureblobstoragezerocopyreplication-gw8-zoo2-1 Removed Stderr: Volume roottestazureblobstoragezerocopyreplication-gw8_data1-1 Removing Stderr: Network roottestazureblobstoragezerocopyreplication-gw8_default Removing Stderr: Volume roottestazureblobstoragezerocopyreplication-gw8_data1-1 Removed Stderr: Network roottestazureblobstoragezerocopyreplication-gw8_default Removed Cleanup called Docker networks for project roottestazureblobstoragezerocopyreplication-gw8 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestazureblobstoragezerocopyreplication-gw8 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestazureblobstoragezerocopyreplication-gw8 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestazureblobstoragezerocopyreplication-gw8-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} No running containers for project: roottestazureblobstoragezerocopyreplication-gw8 Trying to prune unused networks... Trying to prune unused images... Command:[docker image prune -f] Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Executing query select dictGetUInt16('LocalClickHouse_flat_', 'UInt16_', toUInt64(1)) on local_node Stdout:2 Command:[docker volume prune -f] Stdout:Total reclaimed space: 0B Volumes pruned: 2 test_config_corresponding_root/test.py::test_work Running tests in /ClickHouse/tests/integration/test_config_corresponding_root/test.py Cluster start called. is_up=False Docker networks for project roottestconfigcorrespondingroot-gw8 are NETWORK ID NAME DRIVER SCOPE Executing query select dictGetUInt64('MySQL_direct_', 'ParentKeyField', toUInt64(2)) on node Docker containers for project roottestconfigcorrespondingroot-gw8 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestconfigcorrespondingroot-gw8 are DRIVER VOLUME NAME Cleanup called Docker networks for project roottestconfigcorrespondingroot-gw8 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestconfigcorrespondingroot-gw8 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestconfigcorrespondingroot-gw8 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestconfigcorrespondingroot-gw8-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} No running containers for project: roottestconfigcorrespondingroot-gw8 Trying to prune unused networks... Trying to prune unused images... Command:[docker image prune -f] Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Stdout:2 Command:[docker volume prune -f] Stdout:Total reclaimed space: 0B Volumes pruned: 2 Setup directory for instance: node Create directory for configuration generated in this helper Create directory for common tests configuration Copy common configuration from helpers Generate and write macros file Copy custom test config files ['/ClickHouse/tests/integration/test_config_corresponding_root/configs/config.d/bad.xml'] to /ClickHouse/tests/integration/test_config_corresponding_root/_instances-0-gw8/node/configs/config.d Setup database dir /ClickHouse/tests/integration/test_config_corresponding_root/_instances-0-gw8/node/database Setup logs dir /ClickHouse/tests/integration/test_config_corresponding_root/_instances-0-gw8/node/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw'} stored in /ClickHouse/tests/integration/test_config_corresponding_root/_instances-0-gw8/.env Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found http://localhost:None "GET /version HTTP/1.1" 200 826 Command:[docker compose --env-file /ClickHouse/tests/integration/test_config_corresponding_root/_instances-0-gw8/.env --project-name roottestconfigcorrespondingroot-gw8 --file /ClickHouse/tests/integration/test_config_corresponding_root/_instances-0-gw8/node/docker-compose.yml pull] Executing query select dictHas('LocalClickHouse_flat_', toUInt64(1)) on local_node Executing query select dictHas('MySQL_direct_', toUInt64(2)) on node Executing query select dictGetOrDefault('LocalClickHouse_flat_', 'UInt16_', toUInt64(9999999), toUInt16(66)) on local_node Executing query select dictGetOrDefault('MySQL_direct_', 'ParentKeyField', toUInt64(9999999), toUInt64(444)) on node Executing query select dictGetUInt16OrDefault('LocalClickHouse_flat_', 'UInt16_', toUInt64(9999999), toUInt16(66)) on local_node Executing query select dictGetUInt64OrDefault('MySQL_direct_', 'ParentKeyField', toUInt64(9999999), toUInt64(444)) on node Executing query select dictGet('LocalClickHouse_flat_', 'UInt32_', toUInt64(1)) on local_node Executing query select dictGetHierarchy('MySQL_direct_', toUInt64(1)) on node Executing query select dictGetUInt32('LocalClickHouse_flat_', 'UInt32_', toUInt64(1)) on local_node Executing query select dictGetHierarchy('MySQL_direct_', toUInt64(2)) on node Executing query select dictHas('LocalClickHouse_flat_', toUInt64(1)) on local_node Executing query select dictIsIn('MySQL_direct_', toUInt64(1), toUInt64(2)) on node Executing query select dictGetOrDefault('LocalClickHouse_flat_', 'UInt32_', toUInt64(9999999), toUInt32(77)) on local_node Executing query select dictIsIn('MySQL_direct_', toUInt64(2), toUInt64(1)) on node Executing query select dictGetUInt32OrDefault('LocalClickHouse_flat_', 'UInt32_', toUInt64(9999999), toUInt32(77)) on local_node [gw3] PASSED test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_simple[direct] test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_simple[flat] Executing query system reload dictionaries on node Connection dropped: socket connection error: None Executing query select dictGet('LocalClickHouse_flat_', 'UInt64_', toUInt64(1)) on local_node Executing query select dictGet('MySQL_flat_', 'UInt8_', toUInt64(1)) on node Executing query select dictGetUInt64('LocalClickHouse_flat_', 'UInt64_', toUInt64(1)) on local_node Executing query select dictGetUInt8('MySQL_flat_', 'UInt8_', toUInt64(1)) on node Executing query select dictHas('LocalClickHouse_flat_', toUInt64(1)) on local_node Executing query select dictHas('MySQL_flat_', toUInt64(1)) on node Executing query select dictGetOrDefault('LocalClickHouse_flat_', 'UInt64_', toUInt64(9999999), toUInt64(88)) on local_node Executing query select dictGetOrDefault('MySQL_flat_', 'UInt8_', toUInt64(9999999), toUInt8(55)) on node Executing query select dictGetUInt8OrDefault('MySQL_flat_', 'UInt8_', toUInt64(9999999), toUInt8(55)) on node Executing query select dictGetUInt64OrDefault('LocalClickHouse_flat_', 'UInt64_', toUInt64(9999999), toUInt64(88)) on local_node Executing query select dictGet('LocalClickHouse_flat_', 'Int8_', toUInt64(1)) on local_node Executing query select dictGet('MySQL_flat_', 'UInt16_', toUInt64(1)) on node Executing query select dictGetInt8('LocalClickHouse_flat_', 'Int8_', toUInt64(1)) on local_node Executing query select dictGetUInt16('MySQL_flat_', 'UInt16_', toUInt64(1)) on node Executing query select dictHas('LocalClickHouse_flat_', toUInt64(1)) on local_node Executing query select dictHas('MySQL_flat_', toUInt64(1)) on node Executing query select dictGetOrDefault('LocalClickHouse_flat_', 'Int8_', toUInt64(9999999), toInt8(-55)) on local_node Executing query select dictGetOrDefault('MySQL_flat_', 'UInt16_', toUInt64(9999999), toUInt16(66)) on node Executing query select dictGetInt8OrDefault('LocalClickHouse_flat_', 'Int8_', toUInt64(9999999), toInt8(-55)) on local_node Executing query select dictGetUInt16OrDefault('MySQL_flat_', 'UInt16_', toUInt64(9999999), toUInt16(66)) on node Executing query select dictGet('LocalClickHouse_flat_', 'Int16_', toUInt64(1)) on local_node Executing query select dictGet('MySQL_flat_', 'UInt32_', toUInt64(1)) on node Executing query select dictGetInt16('LocalClickHouse_flat_', 'Int16_', toUInt64(1)) on local_node Executing query select dictGetUInt32('MySQL_flat_', 'UInt32_', toUInt64(1)) on node Executing query select dictHas('MySQL_flat_', toUInt64(1)) on node Executing query select dictHas('LocalClickHouse_flat_', toUInt64(1)) on local_node Executing query select dictGetOrDefault('MySQL_flat_', 'UInt32_', toUInt64(9999999), toUInt32(77)) on node Executing query select dictGetOrDefault('LocalClickHouse_flat_', 'Int16_', toUInt64(9999999), toInt16(-66)) on local_node Executing query select dictGetUInt32OrDefault('MySQL_flat_', 'UInt32_', toUInt64(9999999), toUInt32(77)) on node Executing query select dictGetInt16OrDefault('LocalClickHouse_flat_', 'Int16_', toUInt64(9999999), toInt16(-66)) on local_node Executing query select dictGet('MySQL_flat_', 'UInt64_', toUInt64(1)) on node Executing query select dictGet('LocalClickHouse_flat_', 'Int32_', toUInt64(1)) on local_node Executing query select dictGetUInt64('MySQL_flat_', 'UInt64_', toUInt64(1)) on node Executing query select dictGetInt32('LocalClickHouse_flat_', 'Int32_', toUInt64(1)) on local_node Executing query select dictHas('MySQL_flat_', toUInt64(1)) on node Executing query select dictHas('LocalClickHouse_flat_', toUInt64(1)) on local_node Executing query select dictGetOrDefault('MySQL_flat_', 'UInt64_', toUInt64(9999999), toUInt64(88)) on node Executing query select dictGetOrDefault('LocalClickHouse_flat_', 'Int32_', toUInt64(9999999), toInt32(-77)) on local_node Executing query select dictGetUInt64OrDefault('MySQL_flat_', 'UInt64_', toUInt64(9999999), toUInt64(88)) on node Executing query select dictGetInt32OrDefault('LocalClickHouse_flat_', 'Int32_', toUInt64(9999999), toInt32(-77)) on local_node Executing query select dictGet('MySQL_flat_', 'Int8_', toUInt64(1)) on node Executing query select dictGet('LocalClickHouse_flat_', 'Int64_', toUInt64(1)) on local_node Executing query select dictGetInt8('MySQL_flat_', 'Int8_', toUInt64(1)) on node Executing query select dictGetInt64('LocalClickHouse_flat_', 'Int64_', toUInt64(1)) on local_node Executing query select dictHas('LocalClickHouse_flat_', toUInt64(1)) on local_node Executing query select dictHas('MySQL_flat_', toUInt64(1)) on node Executing query select dictGetOrDefault('LocalClickHouse_flat_', 'Int64_', toUInt64(9999999), toInt64(-88)) on local_node Executing query select dictGetOrDefault('MySQL_flat_', 'Int8_', toUInt64(9999999), toInt8(-55)) on node Executing query select dictGetInt64OrDefault('LocalClickHouse_flat_', 'Int64_', toUInt64(9999999), toInt64(-88)) on local_node Executing query select dictGetInt8OrDefault('MySQL_flat_', 'Int8_', toUInt64(9999999), toInt8(-55)) on node Executing query select dictGet('MySQL_flat_', 'Int16_', toUInt64(1)) on node Executing query select dictGet('LocalClickHouse_flat_', 'UUID_', toUInt64(1)) on local_node Executing query select dictGetInt16('MySQL_flat_', 'Int16_', toUInt64(1)) on node Executing query select dictGetUUID('LocalClickHouse_flat_', 'UUID_', toUInt64(1)) on local_node Executing query select dictHas('MySQL_flat_', toUInt64(1)) on node Executing query select dictHas('LocalClickHouse_flat_', toUInt64(1)) on local_node Executing query select dictGetOrDefault('MySQL_flat_', 'Int16_', toUInt64(9999999), toInt16(-66)) on node Executing query select dictGetOrDefault('LocalClickHouse_flat_', 'UUID_', toUInt64(9999999), toUUID('550e8400-0000-0000-0000-000000000000')) on local_node Executing query select dictGetInt16OrDefault('MySQL_flat_', 'Int16_', toUInt64(9999999), toInt16(-66)) on node Executing query select dictGetUUIDOrDefault('LocalClickHouse_flat_', 'UUID_', toUInt64(9999999), toUUID('550e8400-0000-0000-0000-000000000000')) on local_node Executing query select dictGet('MySQL_flat_', 'Int32_', toUInt64(1)) on node Executing query select dictGet('LocalClickHouse_flat_', 'Date_', toUInt64(1)) on local_node Executing query select dictGetInt32('MySQL_flat_', 'Int32_', toUInt64(1)) on node Executing query select dictGetDate('LocalClickHouse_flat_', 'Date_', toUInt64(1)) on local_node Executing query select dictHas('MySQL_flat_', toUInt64(1)) on node Executing query select dictHas('LocalClickHouse_flat_', toUInt64(1)) on local_node Executing query select dictGetOrDefault('MySQL_flat_', 'Int32_', toUInt64(9999999), toInt32(-77)) on node Executing query select dictGetOrDefault('LocalClickHouse_flat_', 'Date_', toUInt64(9999999), toDate('2018-12-30')) on local_node Stderr: node Pulling Stderr: node Pulled ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_config_corresponding_root/_instances-0-gw8/.env --project-name roottestconfigcorrespondingroot-gw8 --file /ClickHouse/tests/integration/test_config_corresponding_root/_instances-0-gw8/node/docker-compose.yml up -d --no-recreate') Command:[docker compose --env-file /ClickHouse/tests/integration/test_config_corresponding_root/_instances-0-gw8/.env --project-name roottestconfigcorrespondingroot-gw8 --file /ClickHouse/tests/integration/test_config_corresponding_root/_instances-0-gw8/node/docker-compose.yml up -d --no-recreate] Executing query select dictGetInt32OrDefault('MySQL_flat_', 'Int32_', toUInt64(9999999), toInt32(-77)) on node Executing query select dictGetDateOrDefault('LocalClickHouse_flat_', 'Date_', toUInt64(9999999), toDate('2018-12-30')) on local_node Stderr: Network roottestconfigcorrespondingroot-gw8_default Creating Stderr: Network roottestconfigcorrespondingroot-gw8_default Created Stderr: Container roottestconfigcorrespondingroot-gw8-node-1 Creating Stderr: Container roottestconfigcorrespondingroot-gw8-node-1 Created Stderr: Container roottestconfigcorrespondingroot-gw8-node-1 Starting Stderr: Container roottestconfigcorrespondingroot-gw8-node-1 Started ClickHouse instance created get_instance_ip instance_name=node http://localhost:None "GET /v1.46/containers/roottestconfigcorrespondingroot-gw8-node-1/json HTTP/1.1" 200 None get_instance_ip instance_name=node http://localhost:None "GET /v1.46/containers/roottestconfigcorrespondingroot-gw8-node-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in node, ip: 172.16.1.2... http://localhost:None "GET /v1.46/containers/roottestconfigcorrespondingroot-gw8-node-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/aacb3c8a56c6791d698107aa9a14306901ebe6e47d78fb7e2fa68d46b2a5368d/json HTTP/1.1" 200 None Executing query select dictGet('MySQL_flat_', 'Int64_', toUInt64(1)) on node http://localhost:None "GET /v1.46/containers/aacb3c8a56c6791d698107aa9a14306901ebe6e47d78fb7e2fa68d46b2a5368d/json HTTP/1.1" 200 None Executing query select dictGet('LocalClickHouse_flat_', 'DateTime_', toUInt64(1)) on local_node http://localhost:None "GET /v1.46/containers/aacb3c8a56c6791d698107aa9a14306901ebe6e47d78fb7e2fa68d46b2a5368d/json HTTP/1.1" 200 None Executing query select dictGetInt64('MySQL_flat_', 'Int64_', toUInt64(1)) on node http://localhost:None "GET /v1.46/containers/aacb3c8a56c6791d698107aa9a14306901ebe6e47d78fb7e2fa68d46b2a5368d/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/aacb3c8a56c6791d698107aa9a14306901ebe6e47d78fb7e2fa68d46b2a5368d/json HTTP/1.1" 200 None Executing query select dictGetDateTime('LocalClickHouse_flat_', 'DateTime_', toUInt64(1)) on local_node http://localhost:None "GET /v1.46/containers/aacb3c8a56c6791d698107aa9a14306901ebe6e47d78fb7e2fa68d46b2a5368d/json HTTP/1.1" 200 None Executing query select dictHas('MySQL_flat_', toUInt64(1)) on node http://localhost:None "GET /v1.46/containers/aacb3c8a56c6791d698107aa9a14306901ebe6e47d78fb7e2fa68d46b2a5368d/json HTTP/1.1" 200 None Executing query select dictHas('LocalClickHouse_flat_', toUInt64(1)) on local_node http://localhost:None "GET /v1.46/containers/aacb3c8a56c6791d698107aa9a14306901ebe6e47d78fb7e2fa68d46b2a5368d/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/aacb3c8a56c6791d698107aa9a14306901ebe6e47d78fb7e2fa68d46b2a5368d/logs?stderr=1&stdout=1×tamps=0&follow=0&tail=all HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/aacb3c8a56c6791d698107aa9a14306901ebe6e47d78fb7e2fa68d46b2a5368d/json HTTP/1.1" 200 None Failed to start cluster: Instance `node' failed to start. Container status: exited, logs: Processing configuration file '/etc/clickhouse-server/config.xml'. Merging configuration file '/etc/clickhouse-server/conf.d/macros.xml'. Merging configuration file '/etc/clickhouse-server/config.d/0_common_enable_keeper_async_replication.xml'. Merging configuration file '/etc/clickhouse-server/config.d/0_common_instance_config.xml'. Merging configuration file '/etc/clickhouse-server/config.d/bad.xml'. Poco::Exception. Code: 1000, e.code() = 0, Exception: Failed to merge config with '/etc/clickhouse-server/config.d/bad.xml': Exception: Root element doesn't have the corresponding root element as the config file. It must be , Stack trace (when copying this message, always include the lines below): 0. ./contrib/llvm-project/libcxx/include/__exception/exception.h:106: Poco::Exception::Exception(String&&, int) @ 0x000000001f44de00 1. ./build_docker/./src/Common/Config/ConfigProcessor.cpp:729: DB::ConfigProcessor::processConfig(bool*, zkutil::ZooKeeperNodeCache*, std::shared_ptr const&, bool) @ 0x000000001b92822d 2. ./build_docker/./src/Common/Config/ConfigProcessor.cpp:802: DB::ConfigProcessor::loadConfig(bool, bool) @ 0x000000001b9285d1 3. ./build_docker/./src/Daemon/BaseDaemon.cpp:126: BaseDaemon::reloadConfiguration() @ 0x000000000fffc510 4. ./build_docker/./src/Daemon/BaseDaemon.cpp:255: BaseDaemon::initialize(Poco::Util::Application&) @ 0x000000000fffddd6 5. ./build_docker/./programs/server/Server.cpp:579: DB::Server::initialize(Poco::Util::Application&) @ 0x000000000ff24ca4 6. ./build_docker/./base/poco/Util/src/Application.cpp:310: Poco::Util::Application::run() @ 0x000000001f592313 7. ./build_docker/./programs/server/Server.cpp:573: DB::Server::run() @ 0x000000000ff24b06 8. ./build_docker/./base/poco/Util/src/ServerApplication.cpp:131: Poco::Util::ServerApplication::run(int, char**) @ 0x000000001f5afd21 9. ./build_docker/./programs/server/Server.cpp:371: mainEntryClickHouseServer(int, char**) @ 0x000000000ff2176d 10. ./build_docker/./programs/main.cpp:295: main @ 0x0000000007b43501 11. ? @ 0x00007f4e02d35d90 12. ? @ 0x00007f4e02d35e40 13. _start @ 0x0000000007a9602e (version 25.2.2.173926.altinityantalya.173926 (official build)) Traceback (most recent call last): File "/ClickHouse/tests/integration/helpers/cluster.py", line 3085, in start instance.wait_for_start(start_timeout) File "/ClickHouse/tests/integration/helpers/cluster.py", line 4386, in wait_for_start raise Exception( Exception: Instance `node' failed to start. Container status: exited, logs: Processing configuration file '/etc/clickhouse-server/config.xml'. Merging configuration file '/etc/clickhouse-server/conf.d/macros.xml'. Merging configuration file '/etc/clickhouse-server/config.d/0_common_enable_keeper_async_replication.xml'. Merging configuration file '/etc/clickhouse-server/config.d/0_common_instance_config.xml'. Merging configuration file '/etc/clickhouse-server/config.d/bad.xml'. Poco::Exception. Code: 1000, e.code() = 0, Exception: Failed to merge config with '/etc/clickhouse-server/config.d/bad.xml': Exception: Root element doesn't have the corresponding root element as the config file. It must be , Stack trace (when copying this message, always include the lines below): 0. ./contrib/llvm-project/libcxx/include/__exception/exception.h:106: Poco::Exception::Exception(String&&, int) @ 0x000000001f44de00 1. ./build_docker/./src/Common/Config/ConfigProcessor.cpp:729: DB::ConfigProcessor::processConfig(bool*, zkutil::ZooKeeperNodeCache*, std::shared_ptr const&, bool) @ 0x000000001b92822d 2. ./build_docker/./src/Common/Config/ConfigProcessor.cpp:802: DB::ConfigProcessor::loadConfig(bool, bool) @ 0x000000001b9285d1 3. ./build_docker/./src/Daemon/BaseDaemon.cpp:126: BaseDaemon::reloadConfiguration() @ 0x000000000fffc510 4. ./build_docker/./src/Daemon/BaseDaemon.cpp:255: BaseDaemon::initialize(Poco::Util::Application&) @ 0x000000000fffddd6 5. ./build_docker/./programs/server/Server.cpp:579: DB::Server::initialize(Poco::Util::Application&) @ 0x000000000ff24ca4 6. ./build_docker/./base/poco/Util/src/Application.cpp:310: Poco::Util::Application::run() @ 0x000000001f592313 7. ./build_docker/./programs/server/Server.cpp:573: DB::Server::run() @ 0x000000000ff24b06 8. ./build_docker/./base/poco/Util/src/ServerApplication.cpp:131: Poco::Util::ServerApplication::run(int, char**) @ 0x000000001f5afd21 9. ./build_docker/./programs/server/Server.cpp:371: mainEntryClickHouseServer(int, char**) @ 0x000000000ff2176d 10. ./build_docker/./programs/main.cpp:295: main @ 0x0000000007b43501 11. ? @ 0x00007f4e02d35d90 12. ? @ 0x00007f4e02d35e40 13. _start @ 0x0000000007a9602e (version 25.2.2.173926.altinityantalya.173926 (official build)) Command:[docker compose --env-file /ClickHouse/tests/integration/test_config_corresponding_root/_instances-0-gw8/.env --project-name roottestconfigcorrespondingroot-gw8 --file /ClickHouse/tests/integration/test_config_corresponding_root/_instances-0-gw8/node/docker-compose.yml stop --timeout 20] Stderr: Container roottestconfigcorrespondingroot-gw8-node-1 Stopping Stderr: Container roottestconfigcorrespondingroot-gw8-node-1 Stopped Command:[bash -c [ -f /ClickHouse/tests/integration/test_config_corresponding_root/_instances-0-gw8/node/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_config_corresponding_root/_instances-0-gw8/node/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[docker compose --env-file /ClickHouse/tests/integration/test_config_corresponding_root/_instances-0-gw8/.env --project-name roottestconfigcorrespondingroot-gw8 --file /ClickHouse/tests/integration/test_config_corresponding_root/_instances-0-gw8/node/docker-compose.yml down --volumes] Executing query select dictGetOrDefault('MySQL_flat_', 'Int64_', toUInt64(9999999), toInt64(-88)) on node Executing query select dictGetOrDefault('LocalClickHouse_flat_', 'DateTime_', toUInt64(9999999), toDateTime('2018-12-30 00:00:00')) on local_node Executing query select dictGetDateTimeOrDefault('LocalClickHouse_flat_', 'DateTime_', toUInt64(9999999), toDateTime('2018-12-30 00:00:00')) on local_node Executing query select dictGetInt64OrDefault('MySQL_flat_', 'Int64_', toUInt64(9999999), toInt64(-88)) on node Executing query select dictGet('LocalClickHouse_flat_', 'String_', toUInt64(1)) on local_node Executing query select dictGet('MySQL_flat_', 'UUID_', toUInt64(1)) on node Stderr: Container roottestconfigcorrespondingroot-gw8-node-1 Stopping Stderr: Container roottestconfigcorrespondingroot-gw8-node-1 Stopped Stderr: Container roottestconfigcorrespondingroot-gw8-node-1 Removing Stderr: Container roottestconfigcorrespondingroot-gw8-node-1 Removed Stderr: Network roottestconfigcorrespondingroot-gw8_default Removing Stderr: Network roottestconfigcorrespondingroot-gw8_default Removed Cleanup called Docker networks for project roottestconfigcorrespondingroot-gw8 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestconfigcorrespondingroot-gw8 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestconfigcorrespondingroot-gw8 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestconfigcorrespondingroot-gw8-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} No running containers for project: roottestconfigcorrespondingroot-gw8 Trying to prune unused networks... Trying to prune unused images... Command:[docker image prune -f] Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Stdout:2 Command:[docker volume prune -f] Stdout:Total reclaimed space: 0B Volumes pruned: 2 [gw8] PASSED test_config_corresponding_root/test.py::test_work Executing query select dictGetString('LocalClickHouse_flat_', 'String_', toUInt64(1)) on local_node Executing query select dictGetUUID('MySQL_flat_', 'UUID_', toUInt64(1)) on node Executing query select dictHas('LocalClickHouse_flat_', toUInt64(1)) on local_node Executing query select dictHas('MySQL_flat_', toUInt64(1)) on node Executing query select dictGetOrDefault('LocalClickHouse_flat_', 'String_', toUInt64(9999999), toString('hi')) on local_node Executing query select dictGetOrDefault('MySQL_flat_', 'UUID_', toUInt64(9999999), toUUID('550e8400-0000-0000-0000-000000000000')) on node Executing query select dictGetStringOrDefault('LocalClickHouse_flat_', 'String_', toUInt64(9999999), toString('hi')) on local_node Executing query select dictGetUUIDOrDefault('MySQL_flat_', 'UUID_', toUInt64(9999999), toUUID('550e8400-0000-0000-0000-000000000000')) on node Executing query select dictGet('LocalClickHouse_flat_', 'Float32_', toUInt64(1)) on local_node Executing query select dictGet('MySQL_flat_', 'Date_', toUInt64(1)) on node Executing query select dictGetFloat32('LocalClickHouse_flat_', 'Float32_', toUInt64(1)) on local_node Executing query select dictGetDate('MySQL_flat_', 'Date_', toUInt64(1)) on node Executing query select dictHas('LocalClickHouse_flat_', toUInt64(1)) on local_node Executing query select dictHas('MySQL_flat_', toUInt64(1)) on node Executing query select dictGetOrDefault('LocalClickHouse_flat_', 'Float32_', toUInt64(9999999), toFloat32(555.11)) on local_node Executing query select dictGetOrDefault('MySQL_flat_', 'Date_', toUInt64(9999999), toDate('2018-12-30')) on node Executing query select dictGetFloat32OrDefault('LocalClickHouse_flat_', 'Float32_', toUInt64(9999999), toFloat32(555.11)) on local_node Executing query select dictGetDateOrDefault('MySQL_flat_', 'Date_', toUInt64(9999999), toDate('2018-12-30')) on node Executing query select dictGet('LocalClickHouse_flat_', 'Float64_', toUInt64(1)) on local_node Executing query select dictGet('MySQL_flat_', 'DateTime_', toUInt64(1)) on node Executing query select dictGetFloat64('LocalClickHouse_flat_', 'Float64_', toUInt64(1)) on local_node Executing query select dictGetDateTime('MySQL_flat_', 'DateTime_', toUInt64(1)) on node Executing query select dictHas('LocalClickHouse_flat_', toUInt64(1)) on local_node Executing query select dictHas('MySQL_flat_', toUInt64(1)) on node Executing query select dictGetOrDefault('LocalClickHouse_flat_', 'Float64_', toUInt64(9999999), toFloat64(777.11)) on local_node Executing query select dictGetOrDefault('MySQL_flat_', 'DateTime_', toUInt64(9999999), toDateTime('2018-12-30 00:00:00')) on node Executing query select dictGetFloat64OrDefault('LocalClickHouse_flat_', 'Float64_', toUInt64(9999999), toFloat64(777.11)) on local_node Executing query select dictGetDateTimeOrDefault('MySQL_flat_', 'DateTime_', toUInt64(9999999), toDateTime('2018-12-30 00:00:00')) on node Executing query select dictGet('LocalClickHouse_flat_', 'ParentKeyField', toUInt64(1)) on local_node Executing query select dictGet('MySQL_flat_', 'String_', toUInt64(1)) on node Executing query select dictGetString('MySQL_flat_', 'String_', toUInt64(1)) on node Executing query select dictGetUInt64('LocalClickHouse_flat_', 'ParentKeyField', toUInt64(1)) on local_node Executing query select dictHas('MySQL_flat_', toUInt64(1)) on node Executing query select dictHas('LocalClickHouse_flat_', toUInt64(1)) on local_node Executing query select dictGetOrDefault('LocalClickHouse_flat_', 'ParentKeyField', toUInt64(9999999), toUInt64(444)) on local_node Executing query select dictGetOrDefault('MySQL_flat_', 'String_', toUInt64(9999999), toString('hi')) on node Executing query select dictGetUInt64OrDefault('LocalClickHouse_flat_', 'ParentKeyField', toUInt64(9999999), toUInt64(444)) on local_node Executing query select dictGetStringOrDefault('MySQL_flat_', 'String_', toUInt64(9999999), toString('hi')) on node Executing query select dictGet('LocalClickHouse_flat_', 'UInt8_', toUInt64(2)) on local_node Executing query select dictGet('MySQL_flat_', 'Float32_', toUInt64(1)) on node Executing query select dictGetUInt8('LocalClickHouse_flat_', 'UInt8_', toUInt64(2)) on local_node Executing query select dictGetFloat32('MySQL_flat_', 'Float32_', toUInt64(1)) on node Executing query select dictHas('LocalClickHouse_flat_', toUInt64(2)) on local_node Executing query select dictHas('MySQL_flat_', toUInt64(1)) on node Executing query select dictGetOrDefault('LocalClickHouse_flat_', 'UInt8_', toUInt64(9999999), toUInt8(55)) on local_node Executing query select dictGetOrDefault('MySQL_flat_', 'Float32_', toUInt64(9999999), toFloat32(555.11)) on node Executing query select dictGetUInt8OrDefault('LocalClickHouse_flat_', 'UInt8_', toUInt64(9999999), toUInt8(55)) on local_node Executing query select dictGetFloat32OrDefault('MySQL_flat_', 'Float32_', toUInt64(9999999), toFloat32(555.11)) on node Executing query select dictGet('LocalClickHouse_flat_', 'UInt16_', toUInt64(2)) on local_node Executing query select dictGet('MySQL_flat_', 'Float64_', toUInt64(1)) on node Executing query select dictGetUInt16('LocalClickHouse_flat_', 'UInt16_', toUInt64(2)) on local_node Executing query select dictGetFloat64('MySQL_flat_', 'Float64_', toUInt64(1)) on node Executing query select dictHas('LocalClickHouse_flat_', toUInt64(2)) on local_node Executing query select dictHas('MySQL_flat_', toUInt64(1)) on node Executing query select dictGetOrDefault('LocalClickHouse_flat_', 'UInt16_', toUInt64(9999999), toUInt16(66)) on local_node Executing query select dictGetOrDefault('MySQL_flat_', 'Float64_', toUInt64(9999999), toFloat64(777.11)) on node Executing query select dictGetUInt16OrDefault('LocalClickHouse_flat_', 'UInt16_', toUInt64(9999999), toUInt16(66)) on local_node Executing query select dictGetFloat64OrDefault('MySQL_flat_', 'Float64_', toUInt64(9999999), toFloat64(777.11)) on node Executing query select dictGet('LocalClickHouse_flat_', 'UInt32_', toUInt64(2)) on local_node Executing query select dictGet('MySQL_flat_', 'ParentKeyField', toUInt64(1)) on node Executing query select dictGetUInt32('LocalClickHouse_flat_', 'UInt32_', toUInt64(2)) on local_node Executing query select dictGetUInt64('MySQL_flat_', 'ParentKeyField', toUInt64(1)) on node Executing query select dictHas('LocalClickHouse_flat_', toUInt64(2)) on local_node Executing query select dictHas('MySQL_flat_', toUInt64(1)) on node Executing query select dictGetOrDefault('LocalClickHouse_flat_', 'UInt32_', toUInt64(9999999), toUInt32(77)) on local_node Executing query select dictGetOrDefault('MySQL_flat_', 'ParentKeyField', toUInt64(9999999), toUInt64(444)) on node Executing query select dictGetUInt32OrDefault('LocalClickHouse_flat_', 'UInt32_', toUInt64(9999999), toUInt32(77)) on local_node Executing query select dictGetUInt64OrDefault('MySQL_flat_', 'ParentKeyField', toUInt64(9999999), toUInt64(444)) on node Executing query select dictGet('LocalClickHouse_flat_', 'UInt64_', toUInt64(2)) on local_node Executing query select dictGet('MySQL_flat_', 'UInt8_', toUInt64(2)) on node Executing query select dictGetUInt64('LocalClickHouse_flat_', 'UInt64_', toUInt64(2)) on local_node Executing query select dictGetUInt8('MySQL_flat_', 'UInt8_', toUInt64(2)) on node Executing query select dictHas('LocalClickHouse_flat_', toUInt64(2)) on local_node Executing query select dictHas('MySQL_flat_', toUInt64(2)) on node Executing query select dictGetOrDefault('LocalClickHouse_flat_', 'UInt64_', toUInt64(9999999), toUInt64(88)) on local_node Executing query select dictGetOrDefault('MySQL_flat_', 'UInt8_', toUInt64(9999999), toUInt8(55)) on node Executing query select dictGetUInt64OrDefault('LocalClickHouse_flat_', 'UInt64_', toUInt64(9999999), toUInt64(88)) on local_node Executing query select dictGetUInt8OrDefault('MySQL_flat_', 'UInt8_', toUInt64(9999999), toUInt8(55)) on node Executing query select dictGet('LocalClickHouse_flat_', 'Int8_', toUInt64(2)) on local_node Executing query select dictGet('MySQL_flat_', 'UInt16_', toUInt64(2)) on node Executing query select dictGetInt8('LocalClickHouse_flat_', 'Int8_', toUInt64(2)) on local_node Executing query select dictGetUInt16('MySQL_flat_', 'UInt16_', toUInt64(2)) on node Executing query select dictHas('LocalClickHouse_flat_', toUInt64(2)) on local_node Executing query select dictHas('MySQL_flat_', toUInt64(2)) on node Executing query select dictGetOrDefault('LocalClickHouse_flat_', 'Int8_', toUInt64(9999999), toInt8(-55)) on local_node Executing query select dictGetOrDefault('MySQL_flat_', 'UInt16_', toUInt64(9999999), toUInt16(66)) on node Executing query select dictGetInt8OrDefault('LocalClickHouse_flat_', 'Int8_', toUInt64(9999999), toInt8(-55)) on local_node Executing query select dictGetUInt16OrDefault('MySQL_flat_', 'UInt16_', toUInt64(9999999), toUInt16(66)) on node Executing query select dictGet('LocalClickHouse_flat_', 'Int16_', toUInt64(2)) on local_node Executing query select dictGet('MySQL_flat_', 'UInt32_', toUInt64(2)) on node Executing query select dictGetInt16('LocalClickHouse_flat_', 'Int16_', toUInt64(2)) on local_node Executing query select dictGetUInt32('MySQL_flat_', 'UInt32_', toUInt64(2)) on node Executing query select dictHas('LocalClickHouse_flat_', toUInt64(2)) on local_node Executing query select dictHas('MySQL_flat_', toUInt64(2)) on node Executing query select dictGetOrDefault('LocalClickHouse_flat_', 'Int16_', toUInt64(9999999), toInt16(-66)) on local_node Executing query select dictGetOrDefault('MySQL_flat_', 'UInt32_', toUInt64(9999999), toUInt32(77)) on node Executing query select dictGetInt16OrDefault('LocalClickHouse_flat_', 'Int16_', toUInt64(9999999), toInt16(-66)) on local_node Executing query select dictGet('LocalClickHouse_flat_', 'Int32_', toUInt64(2)) on local_node Executing query select dictGetUInt32OrDefault('MySQL_flat_', 'UInt32_', toUInt64(9999999), toUInt32(77)) on node Executing query select dictGetInt32('LocalClickHouse_flat_', 'Int32_', toUInt64(2)) on local_node Executing query select dictGet('MySQL_flat_', 'UInt64_', toUInt64(2)) on node Executing query select dictHas('LocalClickHouse_flat_', toUInt64(2)) on local_node Executing query select dictGetUInt64('MySQL_flat_', 'UInt64_', toUInt64(2)) on node Executing query select dictGetOrDefault('LocalClickHouse_flat_', 'Int32_', toUInt64(9999999), toInt32(-77)) on local_node Executing query select dictHas('MySQL_flat_', toUInt64(2)) on node Executing query select dictGetInt32OrDefault('LocalClickHouse_flat_', 'Int32_', toUInt64(9999999), toInt32(-77)) on local_node Executing query select dictGetOrDefault('MySQL_flat_', 'UInt64_', toUInt64(9999999), toUInt64(88)) on node Executing query select dictGet('LocalClickHouse_flat_', 'Int64_', toUInt64(2)) on local_node Executing query select dictGetUInt64OrDefault('MySQL_flat_', 'UInt64_', toUInt64(9999999), toUInt64(88)) on node Executing query select dictGetInt64('LocalClickHouse_flat_', 'Int64_', toUInt64(2)) on local_node Executing query select dictGet('MySQL_flat_', 'Int8_', toUInt64(2)) on node Executing query select dictGetInt8('MySQL_flat_', 'Int8_', toUInt64(2)) on node Executing query select dictHas('LocalClickHouse_flat_', toUInt64(2)) on local_node Executing query select dictHas('MySQL_flat_', toUInt64(2)) on node Executing query select dictGetOrDefault('LocalClickHouse_flat_', 'Int64_', toUInt64(9999999), toInt64(-88)) on local_node Executing query select dictGetOrDefault('MySQL_flat_', 'Int8_', toUInt64(9999999), toInt8(-55)) on node Executing query select dictGetInt64OrDefault('LocalClickHouse_flat_', 'Int64_', toUInt64(9999999), toInt64(-88)) on local_node Executing query select dictGet('LocalClickHouse_flat_', 'UUID_', toUInt64(2)) on local_node Executing query select dictGetInt8OrDefault('MySQL_flat_', 'Int8_', toUInt64(9999999), toInt8(-55)) on node Executing query select dictGetUUID('LocalClickHouse_flat_', 'UUID_', toUInt64(2)) on local_node Executing query select dictGet('MySQL_flat_', 'Int16_', toUInt64(2)) on node Executing query select dictHas('LocalClickHouse_flat_', toUInt64(2)) on local_node Executing query select dictGetInt16('MySQL_flat_', 'Int16_', toUInt64(2)) on node Executing query select dictGetOrDefault('LocalClickHouse_flat_', 'UUID_', toUInt64(9999999), toUUID('550e8400-0000-0000-0000-000000000000')) on local_node Executing query select dictHas('MySQL_flat_', toUInt64(2)) on node Executing query select dictGetUUIDOrDefault('LocalClickHouse_flat_', 'UUID_', toUInt64(9999999), toUUID('550e8400-0000-0000-0000-000000000000')) on local_node Executing query select dictGetOrDefault('MySQL_flat_', 'Int16_', toUInt64(9999999), toInt16(-66)) on node Executing query select dictGetInt16OrDefault('MySQL_flat_', 'Int16_', toUInt64(9999999), toInt16(-66)) on node Executing query select dictGet('LocalClickHouse_flat_', 'Date_', toUInt64(2)) on local_node Executing query select dictGetDate('LocalClickHouse_flat_', 'Date_', toUInt64(2)) on local_node Executing query select dictGet('MySQL_flat_', 'Int32_', toUInt64(2)) on node Executing query select dictHas('LocalClickHouse_flat_', toUInt64(2)) on local_node Executing query select dictGetInt32('MySQL_flat_', 'Int32_', toUInt64(2)) on node Executing query select dictHas('MySQL_flat_', toUInt64(2)) on node Executing query select dictGetOrDefault('LocalClickHouse_flat_', 'Date_', toUInt64(9999999), toDate('2018-12-30')) on local_node Executing query select dictGetOrDefault('MySQL_flat_', 'Int32_', toUInt64(9999999), toInt32(-77)) on node Executing query select dictGetDateOrDefault('LocalClickHouse_flat_', 'Date_', toUInt64(9999999), toDate('2018-12-30')) on local_node Executing query select dictGetInt32OrDefault('MySQL_flat_', 'Int32_', toUInt64(9999999), toInt32(-77)) on node Executing query select dictGet('LocalClickHouse_flat_', 'DateTime_', toUInt64(2)) on local_node Executing query select dictGet('MySQL_flat_', 'Int64_', toUInt64(2)) on node Executing query select dictGetDateTime('LocalClickHouse_flat_', 'DateTime_', toUInt64(2)) on local_node Executing query select dictHas('LocalClickHouse_flat_', toUInt64(2)) on local_node Executing query select dictGetInt64('MySQL_flat_', 'Int64_', toUInt64(2)) on node Executing query select dictGetOrDefault('LocalClickHouse_flat_', 'DateTime_', toUInt64(9999999), toDateTime('2018-12-30 00:00:00')) on local_node Executing query select dictHas('MySQL_flat_', toUInt64(2)) on node Executing query select dictGetDateTimeOrDefault('LocalClickHouse_flat_', 'DateTime_', toUInt64(9999999), toDateTime('2018-12-30 00:00:00')) on local_node Executing query select dictGetOrDefault('MySQL_flat_', 'Int64_', toUInt64(9999999), toInt64(-88)) on node Executing query select dictGet('LocalClickHouse_flat_', 'String_', toUInt64(2)) on local_node Executing query select dictGetInt64OrDefault('MySQL_flat_', 'Int64_', toUInt64(9999999), toInt64(-88)) on node Connection dropped: socket connection error: None Executing query select dictGetString('LocalClickHouse_flat_', 'String_', toUInt64(2)) on local_node Executing query select dictGet('MySQL_flat_', 'UUID_', toUInt64(2)) on node Executing query select dictHas('LocalClickHouse_flat_', toUInt64(2)) on local_node Executing query select dictGetUUID('MySQL_flat_', 'UUID_', toUInt64(2)) on node Executing query select dictGetOrDefault('LocalClickHouse_flat_', 'String_', toUInt64(9999999), toString('hi')) on local_node Executing query select dictHas('MySQL_flat_', toUInt64(2)) on node Executing query select dictGetStringOrDefault('LocalClickHouse_flat_', 'String_', toUInt64(9999999), toString('hi')) on local_node Executing query select dictGetOrDefault('MySQL_flat_', 'UUID_', toUInt64(9999999), toUUID('550e8400-0000-0000-0000-000000000000')) on node Executing query select dictGet('LocalClickHouse_flat_', 'Float32_', toUInt64(2)) on local_node Executing query select dictGetUUIDOrDefault('MySQL_flat_', 'UUID_', toUInt64(9999999), toUUID('550e8400-0000-0000-0000-000000000000')) on node Executing query select dictGetFloat32('LocalClickHouse_flat_', 'Float32_', toUInt64(2)) on local_node Executing query select dictGet('MySQL_flat_', 'Date_', toUInt64(2)) on node Executing query select dictHas('LocalClickHouse_flat_', toUInt64(2)) on local_node Executing query select dictGetDate('MySQL_flat_', 'Date_', toUInt64(2)) on node Executing query select dictGetOrDefault('LocalClickHouse_flat_', 'Float32_', toUInt64(9999999), toFloat32(555.11)) on local_node Executing query select dictHas('MySQL_flat_', toUInt64(2)) on node Executing query select dictGetFloat32OrDefault('LocalClickHouse_flat_', 'Float32_', toUInt64(9999999), toFloat32(555.11)) on local_node Executing query select dictGetOrDefault('MySQL_flat_', 'Date_', toUInt64(9999999), toDate('2018-12-30')) on node Executing query select dictGet('LocalClickHouse_flat_', 'Float64_', toUInt64(2)) on local_node Executing query select dictGetDateOrDefault('MySQL_flat_', 'Date_', toUInt64(9999999), toDate('2018-12-30')) on node Executing query select dictGetFloat64('LocalClickHouse_flat_', 'Float64_', toUInt64(2)) on local_node Executing query select dictGet('MySQL_flat_', 'DateTime_', toUInt64(2)) on node Executing query select dictHas('LocalClickHouse_flat_', toUInt64(2)) on local_node Executing query select dictGetDateTime('MySQL_flat_', 'DateTime_', toUInt64(2)) on node Executing query select dictGetOrDefault('LocalClickHouse_flat_', 'Float64_', toUInt64(9999999), toFloat64(777.11)) on local_node Executing query select dictHas('MySQL_flat_', toUInt64(2)) on node Executing query select dictGetFloat64OrDefault('LocalClickHouse_flat_', 'Float64_', toUInt64(9999999), toFloat64(777.11)) on local_node Executing query select dictGetOrDefault('MySQL_flat_', 'DateTime_', toUInt64(9999999), toDateTime('2018-12-30 00:00:00')) on node Executing query select dictGet('LocalClickHouse_flat_', 'ParentKeyField', toUInt64(2)) on local_node Executing query select dictGetDateTimeOrDefault('MySQL_flat_', 'DateTime_', toUInt64(9999999), toDateTime('2018-12-30 00:00:00')) on node Executing query select dictGetUInt64('LocalClickHouse_flat_', 'ParentKeyField', toUInt64(2)) on local_node Executing query select dictGet('MySQL_flat_', 'String_', toUInt64(2)) on node Executing query select dictHas('LocalClickHouse_flat_', toUInt64(2)) on local_node Executing query select dictGetString('MySQL_flat_', 'String_', toUInt64(2)) on node Executing query select dictGetOrDefault('LocalClickHouse_flat_', 'ParentKeyField', toUInt64(9999999), toUInt64(444)) on local_node Executing query select dictHas('MySQL_flat_', toUInt64(2)) on node Executing query select dictGetUInt64OrDefault('LocalClickHouse_flat_', 'ParentKeyField', toUInt64(9999999), toUInt64(444)) on local_node Executing query select dictGetOrDefault('MySQL_flat_', 'String_', toUInt64(9999999), toString('hi')) on node Executing query select dictGetHierarchy('LocalClickHouse_flat_', toUInt64(1)) on local_node Executing query select dictGetStringOrDefault('MySQL_flat_', 'String_', toUInt64(9999999), toString('hi')) on node Executing query select dictGetHierarchy('LocalClickHouse_flat_', toUInt64(2)) on local_node Executing query select dictGet('MySQL_flat_', 'Float32_', toUInt64(2)) on node Executing query select dictIsIn('LocalClickHouse_flat_', toUInt64(1), toUInt64(2)) on local_node Executing query select dictGetFloat32('MySQL_flat_', 'Float32_', toUInt64(2)) on node Executing query select dictIsIn('LocalClickHouse_flat_', toUInt64(2), toUInt64(1)) on local_node Executing query select dictHas('MySQL_flat_', toUInt64(2)) on node [gw5] PASSED test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_simple[flat] test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_simple[hashed] Executing query system reload dictionaries on local_node Executing query select dictGetOrDefault('MySQL_flat_', 'Float32_', toUInt64(9999999), toFloat32(555.11)) on node Executing query select dictGet('LocalClickHouse_hashed_', 'UInt8_', toUInt64(1)) on local_node Executing query select dictGetFloat32OrDefault('MySQL_flat_', 'Float32_', toUInt64(9999999), toFloat32(555.11)) on node Executing query select dictGetUInt8('LocalClickHouse_hashed_', 'UInt8_', toUInt64(1)) on local_node Executing query select dictGet('MySQL_flat_', 'Float64_', toUInt64(2)) on node Executing query select dictHas('LocalClickHouse_hashed_', toUInt64(1)) on local_node Executing query select dictGetFloat64('MySQL_flat_', 'Float64_', toUInt64(2)) on node Executing query select dictGetOrDefault('LocalClickHouse_hashed_', 'UInt8_', toUInt64(9999999), toUInt8(55)) on local_node Executing query select dictHas('MySQL_flat_', toUInt64(2)) on node Executing query select dictGetUInt8OrDefault('LocalClickHouse_hashed_', 'UInt8_', toUInt64(9999999), toUInt8(55)) on local_node Executing query select dictGetOrDefault('MySQL_flat_', 'Float64_', toUInt64(9999999), toFloat64(777.11)) on node Executing query select dictGet('LocalClickHouse_hashed_', 'UInt16_', toUInt64(1)) on local_node Executing query select dictGetFloat64OrDefault('MySQL_flat_', 'Float64_', toUInt64(9999999), toFloat64(777.11)) on node Executing query select dictGetUInt16('LocalClickHouse_hashed_', 'UInt16_', toUInt64(1)) on local_node Executing query select dictGet('MySQL_flat_', 'ParentKeyField', toUInt64(2)) on node Executing query select dictHas('LocalClickHouse_hashed_', toUInt64(1)) on local_node Executing query select dictGetUInt64('MySQL_flat_', 'ParentKeyField', toUInt64(2)) on node Executing query select dictGetOrDefault('LocalClickHouse_hashed_', 'UInt16_', toUInt64(9999999), toUInt16(66)) on local_node Executing query select dictHas('MySQL_flat_', toUInt64(2)) on node Executing query select dictGetUInt16OrDefault('LocalClickHouse_hashed_', 'UInt16_', toUInt64(9999999), toUInt16(66)) on local_node Executing query select dictGetOrDefault('MySQL_flat_', 'ParentKeyField', toUInt64(9999999), toUInt64(444)) on node Executing query select dictGet('LocalClickHouse_hashed_', 'UInt32_', toUInt64(1)) on local_node Executing query select dictGetUInt64OrDefault('MySQL_flat_', 'ParentKeyField', toUInt64(9999999), toUInt64(444)) on node Executing query select dictGetUInt32('LocalClickHouse_hashed_', 'UInt32_', toUInt64(1)) on local_node Executing query select dictGetHierarchy('MySQL_flat_', toUInt64(1)) on node Executing query select dictHas('LocalClickHouse_hashed_', toUInt64(1)) on local_node Executing query select dictGetHierarchy('MySQL_flat_', toUInt64(2)) on node Executing query select dictGetOrDefault('LocalClickHouse_hashed_', 'UInt32_', toUInt64(9999999), toUInt32(77)) on local_node Executing query select dictIsIn('MySQL_flat_', toUInt64(1), toUInt64(2)) on node Executing query select dictGetUInt32OrDefault('LocalClickHouse_hashed_', 'UInt32_', toUInt64(9999999), toUInt32(77)) on local_node Executing query select dictIsIn('MySQL_flat_', toUInt64(2), toUInt64(1)) on node Executing query select dictGet('LocalClickHouse_hashed_', 'UInt64_', toUInt64(1)) on local_node [gw3] PASSED test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_simple[flat] test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_simple[hashed] Executing query system reload dictionaries on node Executing query select dictGetUInt64('LocalClickHouse_hashed_', 'UInt64_', toUInt64(1)) on local_node Executing query select dictGet('MySQL_hashed_', 'UInt8_', toUInt64(1)) on node Executing query select dictHas('LocalClickHouse_hashed_', toUInt64(1)) on local_node Executing query select dictGetUInt8('MySQL_hashed_', 'UInt8_', toUInt64(1)) on node Executing query select dictGetOrDefault('LocalClickHouse_hashed_', 'UInt64_', toUInt64(9999999), toUInt64(88)) on local_node Executing query select dictHas('MySQL_hashed_', toUInt64(1)) on node Executing query select dictGetUInt64OrDefault('LocalClickHouse_hashed_', 'UInt64_', toUInt64(9999999), toUInt64(88)) on local_node Executing query select dictGetOrDefault('MySQL_hashed_', 'UInt8_', toUInt64(9999999), toUInt8(55)) on node Executing query select dictGet('LocalClickHouse_hashed_', 'Int8_', toUInt64(1)) on local_node Executing query select dictGetUInt8OrDefault('MySQL_hashed_', 'UInt8_', toUInt64(9999999), toUInt8(55)) on node Executing query select dictGetInt8('LocalClickHouse_hashed_', 'Int8_', toUInt64(1)) on local_node Executing query select dictGet('MySQL_hashed_', 'UInt16_', toUInt64(1)) on node Executing query select dictHas('LocalClickHouse_hashed_', toUInt64(1)) on local_node Executing query select dictGetUInt16('MySQL_hashed_', 'UInt16_', toUInt64(1)) on node Executing query select dictGetOrDefault('LocalClickHouse_hashed_', 'Int8_', toUInt64(9999999), toInt8(-55)) on local_node Executing query select dictHas('MySQL_hashed_', toUInt64(1)) on node Executing query select dictGetInt8OrDefault('LocalClickHouse_hashed_', 'Int8_', toUInt64(9999999), toInt8(-55)) on local_node Executing query select dictGetOrDefault('MySQL_hashed_', 'UInt16_', toUInt64(9999999), toUInt16(66)) on node Executing query select dictGet('LocalClickHouse_hashed_', 'Int16_', toUInt64(1)) on local_node Executing query select dictGetUInt16OrDefault('MySQL_hashed_', 'UInt16_', toUInt64(9999999), toUInt16(66)) on node Executing query select dictGetInt16('LocalClickHouse_hashed_', 'Int16_', toUInt64(1)) on local_node Executing query select dictGet('MySQL_hashed_', 'UInt32_', toUInt64(1)) on node Executing query select dictHas('LocalClickHouse_hashed_', toUInt64(1)) on local_node Executing query select dictGetUInt32('MySQL_hashed_', 'UInt32_', toUInt64(1)) on node Executing query select dictGetOrDefault('LocalClickHouse_hashed_', 'Int16_', toUInt64(9999999), toInt16(-66)) on local_node Executing query select dictHas('MySQL_hashed_', toUInt64(1)) on node Executing query select dictGetOrDefault('MySQL_hashed_', 'UInt32_', toUInt64(9999999), toUInt32(77)) on node Executing query select dictGetInt16OrDefault('LocalClickHouse_hashed_', 'Int16_', toUInt64(9999999), toInt16(-66)) on local_node Executing query select dictGetUInt32OrDefault('MySQL_hashed_', 'UInt32_', toUInt64(9999999), toUInt32(77)) on node Executing query select dictGet('LocalClickHouse_hashed_', 'Int32_', toUInt64(1)) on local_node Executing query select dictGet('MySQL_hashed_', 'UInt64_', toUInt64(1)) on node Executing query select dictGetInt32('LocalClickHouse_hashed_', 'Int32_', toUInt64(1)) on local_node Executing query select dictGetUInt64('MySQL_hashed_', 'UInt64_', toUInt64(1)) on node Executing query select dictHas('LocalClickHouse_hashed_', toUInt64(1)) on local_node Executing query select dictHas('MySQL_hashed_', toUInt64(1)) on node Executing query select dictGetOrDefault('LocalClickHouse_hashed_', 'Int32_', toUInt64(9999999), toInt32(-77)) on local_node Executing query select dictGetOrDefault('MySQL_hashed_', 'UInt64_', toUInt64(9999999), toUInt64(88)) on node Executing query select dictGetInt32OrDefault('LocalClickHouse_hashed_', 'Int32_', toUInt64(9999999), toInt32(-77)) on local_node Executing query select dictGetUInt64OrDefault('MySQL_hashed_', 'UInt64_', toUInt64(9999999), toUInt64(88)) on node Executing query select dictGet('LocalClickHouse_hashed_', 'Int64_', toUInt64(1)) on local_node Executing query select dictGet('MySQL_hashed_', 'Int8_', toUInt64(1)) on node Executing query select dictGetInt64('LocalClickHouse_hashed_', 'Int64_', toUInt64(1)) on local_node Executing query select dictGetInt8('MySQL_hashed_', 'Int8_', toUInt64(1)) on node Executing query select dictHas('LocalClickHouse_hashed_', toUInt64(1)) on local_node Executing query select dictHas('MySQL_hashed_', toUInt64(1)) on node Executing query select dictGetOrDefault('LocalClickHouse_hashed_', 'Int64_', toUInt64(9999999), toInt64(-88)) on local_node Executing query select dictGetOrDefault('MySQL_hashed_', 'Int8_', toUInt64(9999999), toInt8(-55)) on node Executing query select dictGetInt64OrDefault('LocalClickHouse_hashed_', 'Int64_', toUInt64(9999999), toInt64(-88)) on local_node Executing query select dictGetInt8OrDefault('MySQL_hashed_', 'Int8_', toUInt64(9999999), toInt8(-55)) on node Executing query select dictGet('LocalClickHouse_hashed_', 'UUID_', toUInt64(1)) on local_node Executing query select dictGet('MySQL_hashed_', 'Int16_', toUInt64(1)) on node Executing query select dictGetUUID('LocalClickHouse_hashed_', 'UUID_', toUInt64(1)) on local_node Executing query select dictGetInt16('MySQL_hashed_', 'Int16_', toUInt64(1)) on node Executing query select dictHas('LocalClickHouse_hashed_', toUInt64(1)) on local_node Executing query select dictHas('MySQL_hashed_', toUInt64(1)) on node Executing query select dictGetOrDefault('LocalClickHouse_hashed_', 'UUID_', toUInt64(9999999), toUUID('550e8400-0000-0000-0000-000000000000')) on local_node Executing query select dictGetUUIDOrDefault('LocalClickHouse_hashed_', 'UUID_', toUInt64(9999999), toUUID('550e8400-0000-0000-0000-000000000000')) on local_node Executing query select dictGetOrDefault('MySQL_hashed_', 'Int16_', toUInt64(9999999), toInt16(-66)) on node Executing query select dictGetInt16OrDefault('MySQL_hashed_', 'Int16_', toUInt64(9999999), toInt16(-66)) on node Executing query select dictGet('LocalClickHouse_hashed_', 'Date_', toUInt64(1)) on local_node Executing query select dictGet('MySQL_hashed_', 'Int32_', toUInt64(1)) on node Executing query select dictGetDate('LocalClickHouse_hashed_', 'Date_', toUInt64(1)) on local_node Executing query select dictGetInt32('MySQL_hashed_', 'Int32_', toUInt64(1)) on node Executing query select dictHas('LocalClickHouse_hashed_', toUInt64(1)) on local_node Executing query select dictHas('MySQL_hashed_', toUInt64(1)) on node Executing query select dictGetOrDefault('LocalClickHouse_hashed_', 'Date_', toUInt64(9999999), toDate('2018-12-30')) on local_node Executing query select dictGetOrDefault('MySQL_hashed_', 'Int32_', toUInt64(9999999), toInt32(-77)) on node Executing query select dictGetDateOrDefault('LocalClickHouse_hashed_', 'Date_', toUInt64(9999999), toDate('2018-12-30')) on local_node Executing query select dictGetInt32OrDefault('MySQL_hashed_', 'Int32_', toUInt64(9999999), toInt32(-77)) on node Executing query select dictGet('LocalClickHouse_hashed_', 'DateTime_', toUInt64(1)) on local_node Executing query select dictGet('MySQL_hashed_', 'Int64_', toUInt64(1)) on node Executing query select dictGetDateTime('LocalClickHouse_hashed_', 'DateTime_', toUInt64(1)) on local_node Executing query select dictGetInt64('MySQL_hashed_', 'Int64_', toUInt64(1)) on node Executing query select dictHas('LocalClickHouse_hashed_', toUInt64(1)) on local_node Executing query select dictHas('MySQL_hashed_', toUInt64(1)) on node Executing query select dictGetOrDefault('LocalClickHouse_hashed_', 'DateTime_', toUInt64(9999999), toDateTime('2018-12-30 00:00:00')) on local_node Executing query select dictGetOrDefault('MySQL_hashed_', 'Int64_', toUInt64(9999999), toInt64(-88)) on node Executing query select dictGetDateTimeOrDefault('LocalClickHouse_hashed_', 'DateTime_', toUInt64(9999999), toDateTime('2018-12-30 00:00:00')) on local_node Executing query select dictGetInt64OrDefault('MySQL_hashed_', 'Int64_', toUInt64(9999999), toInt64(-88)) on node Executing query select dictGet('LocalClickHouse_hashed_', 'String_', toUInt64(1)) on local_node Executing query select dictGet('MySQL_hashed_', 'UUID_', toUInt64(1)) on node Executing query select dictGetString('LocalClickHouse_hashed_', 'String_', toUInt64(1)) on local_node Executing query select dictGetUUID('MySQL_hashed_', 'UUID_', toUInt64(1)) on node Executing query select dictHas('LocalClickHouse_hashed_', toUInt64(1)) on local_node Executing query select dictHas('MySQL_hashed_', toUInt64(1)) on node Executing query select dictGetOrDefault('LocalClickHouse_hashed_', 'String_', toUInt64(9999999), toString('hi')) on local_node Executing query select dictGetOrDefault('MySQL_hashed_', 'UUID_', toUInt64(9999999), toUUID('550e8400-0000-0000-0000-000000000000')) on node Executing query select dictGetStringOrDefault('LocalClickHouse_hashed_', 'String_', toUInt64(9999999), toString('hi')) on local_node Executing query select dictGetUUIDOrDefault('MySQL_hashed_', 'UUID_', toUInt64(9999999), toUUID('550e8400-0000-0000-0000-000000000000')) on node Executing query select dictGet('LocalClickHouse_hashed_', 'Float32_', toUInt64(1)) on local_node Executing query select dictGet('MySQL_hashed_', 'Date_', toUInt64(1)) on node Executing query select dictGetFloat32('LocalClickHouse_hashed_', 'Float32_', toUInt64(1)) on local_node Executing query select dictGetDate('MySQL_hashed_', 'Date_', toUInt64(1)) on node Executing query select dictHas('LocalClickHouse_hashed_', toUInt64(1)) on local_node Executing query select dictHas('MySQL_hashed_', toUInt64(1)) on node Executing query select dictGetOrDefault('LocalClickHouse_hashed_', 'Float32_', toUInt64(9999999), toFloat32(555.11)) on local_node Executing query select dictGetOrDefault('MySQL_hashed_', 'Date_', toUInt64(9999999), toDate('2018-12-30')) on node Executing query select dictGetFloat32OrDefault('LocalClickHouse_hashed_', 'Float32_', toUInt64(9999999), toFloat32(555.11)) on local_node Executing query select dictGetDateOrDefault('MySQL_hashed_', 'Date_', toUInt64(9999999), toDate('2018-12-30')) on node Executing query select dictGet('LocalClickHouse_hashed_', 'Float64_', toUInt64(1)) on local_node Executing query select dictGet('MySQL_hashed_', 'DateTime_', toUInt64(1)) on node Executing query select dictGetFloat64('LocalClickHouse_hashed_', 'Float64_', toUInt64(1)) on local_node Executing query select dictGetDateTime('MySQL_hashed_', 'DateTime_', toUInt64(1)) on node Executing query select dictHas('LocalClickHouse_hashed_', toUInt64(1)) on local_node Executing query select dictHas('MySQL_hashed_', toUInt64(1)) on node Executing query select dictGetOrDefault('LocalClickHouse_hashed_', 'Float64_', toUInt64(9999999), toFloat64(777.11)) on local_node Executing query select dictGetOrDefault('MySQL_hashed_', 'DateTime_', toUInt64(9999999), toDateTime('2018-12-30 00:00:00')) on node Executing query select dictGetFloat64OrDefault('LocalClickHouse_hashed_', 'Float64_', toUInt64(9999999), toFloat64(777.11)) on local_node Connection dropped: socket connection error: None Executing query select dictGetDateTimeOrDefault('MySQL_hashed_', 'DateTime_', toUInt64(9999999), toDateTime('2018-12-30 00:00:00')) on node Executing query select dictGet('LocalClickHouse_hashed_', 'ParentKeyField', toUInt64(1)) on local_node Executing query select dictGet('MySQL_hashed_', 'String_', toUInt64(1)) on node Executing query select dictGetUInt64('LocalClickHouse_hashed_', 'ParentKeyField', toUInt64(1)) on local_node Executing query select dictGetString('MySQL_hashed_', 'String_', toUInt64(1)) on node Executing query select dictHas('LocalClickHouse_hashed_', toUInt64(1)) on local_node Executing query select dictHas('MySQL_hashed_', toUInt64(1)) on node Executing query select dictGetOrDefault('LocalClickHouse_hashed_', 'ParentKeyField', toUInt64(9999999), toUInt64(444)) on local_node Executing query select dictGetOrDefault('MySQL_hashed_', 'String_', toUInt64(9999999), toString('hi')) on node Executing query select dictGetUInt64OrDefault('LocalClickHouse_hashed_', 'ParentKeyField', toUInt64(9999999), toUInt64(444)) on local_node Executing query select dictGetStringOrDefault('MySQL_hashed_', 'String_', toUInt64(9999999), toString('hi')) on node Executing query select dictGet('LocalClickHouse_hashed_', 'UInt8_', toUInt64(2)) on local_node Executing query select dictGet('MySQL_hashed_', 'Float32_', toUInt64(1)) on node Executing query select dictGetUInt8('LocalClickHouse_hashed_', 'UInt8_', toUInt64(2)) on local_node Executing query select dictGetFloat32('MySQL_hashed_', 'Float32_', toUInt64(1)) on node Executing query select dictHas('LocalClickHouse_hashed_', toUInt64(2)) on local_node Executing query select dictHas('MySQL_hashed_', toUInt64(1)) on node Executing query select dictGetOrDefault('LocalClickHouse_hashed_', 'UInt8_', toUInt64(9999999), toUInt8(55)) on local_node Executing query select dictGetOrDefault('MySQL_hashed_', 'Float32_', toUInt64(9999999), toFloat32(555.11)) on node Executing query select dictGetUInt8OrDefault('LocalClickHouse_hashed_', 'UInt8_', toUInt64(9999999), toUInt8(55)) on local_node Executing query select dictGetFloat32OrDefault('MySQL_hashed_', 'Float32_', toUInt64(9999999), toFloat32(555.11)) on node Executing query select dictGet('LocalClickHouse_hashed_', 'UInt16_', toUInt64(2)) on local_node Executing query select dictGet('MySQL_hashed_', 'Float64_', toUInt64(1)) on node Executing query select dictGetUInt16('LocalClickHouse_hashed_', 'UInt16_', toUInt64(2)) on local_node Executing query select dictGetFloat64('MySQL_hashed_', 'Float64_', toUInt64(1)) on node Executing query select dictHas('LocalClickHouse_hashed_', toUInt64(2)) on local_node Executing query select dictHas('MySQL_hashed_', toUInt64(1)) on node Executing query select dictGetOrDefault('LocalClickHouse_hashed_', 'UInt16_', toUInt64(9999999), toUInt16(66)) on local_node Executing query select dictGetOrDefault('MySQL_hashed_', 'Float64_', toUInt64(9999999), toFloat64(777.11)) on node Executing query select dictGetUInt16OrDefault('LocalClickHouse_hashed_', 'UInt16_', toUInt64(9999999), toUInt16(66)) on local_node Executing query select dictGetFloat64OrDefault('MySQL_hashed_', 'Float64_', toUInt64(9999999), toFloat64(777.11)) on node Executing query select dictGet('LocalClickHouse_hashed_', 'UInt32_', toUInt64(2)) on local_node Executing query select dictGet('MySQL_hashed_', 'ParentKeyField', toUInt64(1)) on node Executing query select dictGetUInt32('LocalClickHouse_hashed_', 'UInt32_', toUInt64(2)) on local_node Executing query select dictGetUInt64('MySQL_hashed_', 'ParentKeyField', toUInt64(1)) on node Executing query select dictHas('LocalClickHouse_hashed_', toUInt64(2)) on local_node Executing query select dictHas('MySQL_hashed_', toUInt64(1)) on node Executing query select dictGetOrDefault('LocalClickHouse_hashed_', 'UInt32_', toUInt64(9999999), toUInt32(77)) on local_node Executing query select dictGetOrDefault('MySQL_hashed_', 'ParentKeyField', toUInt64(9999999), toUInt64(444)) on node Executing query select dictGetUInt32OrDefault('LocalClickHouse_hashed_', 'UInt32_', toUInt64(9999999), toUInt32(77)) on local_node Executing query select dictGetUInt64OrDefault('MySQL_hashed_', 'ParentKeyField', toUInt64(9999999), toUInt64(444)) on node Executing query select dictGet('LocalClickHouse_hashed_', 'UInt64_', toUInt64(2)) on local_node Executing query select dictGet('MySQL_hashed_', 'UInt8_', toUInt64(2)) on node Executing query select dictGetUInt64('LocalClickHouse_hashed_', 'UInt64_', toUInt64(2)) on local_node Executing query select dictGetUInt8('MySQL_hashed_', 'UInt8_', toUInt64(2)) on node Executing query select dictHas('LocalClickHouse_hashed_', toUInt64(2)) on local_node Executing query select dictHas('MySQL_hashed_', toUInt64(2)) on node Executing query select dictGetOrDefault('LocalClickHouse_hashed_', 'UInt64_', toUInt64(9999999), toUInt64(88)) on local_node Executing query select dictGetOrDefault('MySQL_hashed_', 'UInt8_', toUInt64(9999999), toUInt8(55)) on node Executing query select dictGetUInt8OrDefault('MySQL_hashed_', 'UInt8_', toUInt64(9999999), toUInt8(55)) on node Executing query select dictGetUInt64OrDefault('LocalClickHouse_hashed_', 'UInt64_', toUInt64(9999999), toUInt64(88)) on local_node Executing query select dictGet('LocalClickHouse_hashed_', 'Int8_', toUInt64(2)) on local_node Executing query select dictGet('MySQL_hashed_', 'UInt16_', toUInt64(2)) on node Executing query select dictGetUInt16('MySQL_hashed_', 'UInt16_', toUInt64(2)) on node Executing query select dictGetInt8('LocalClickHouse_hashed_', 'Int8_', toUInt64(2)) on local_node Executing query select dictHas('MySQL_hashed_', toUInt64(2)) on node Executing query select dictHas('LocalClickHouse_hashed_', toUInt64(2)) on local_node Executing query select dictGetOrDefault('MySQL_hashed_', 'UInt16_', toUInt64(9999999), toUInt16(66)) on node Executing query select dictGetOrDefault('LocalClickHouse_hashed_', 'Int8_', toUInt64(9999999), toInt8(-55)) on local_node Executing query select dictGetUInt16OrDefault('MySQL_hashed_', 'UInt16_', toUInt64(9999999), toUInt16(66)) on node Executing query select dictGetInt8OrDefault('LocalClickHouse_hashed_', 'Int8_', toUInt64(9999999), toInt8(-55)) on local_node Executing query select dictGet('MySQL_hashed_', 'UInt32_', toUInt64(2)) on node Executing query select dictGet('LocalClickHouse_hashed_', 'Int16_', toUInt64(2)) on local_node Executing query select dictGetUInt32('MySQL_hashed_', 'UInt32_', toUInt64(2)) on node Executing query select dictGetInt16('LocalClickHouse_hashed_', 'Int16_', toUInt64(2)) on local_node Executing query select dictHas('MySQL_hashed_', toUInt64(2)) on node Executing query select dictHas('LocalClickHouse_hashed_', toUInt64(2)) on local_node Executing query select dictGetOrDefault('MySQL_hashed_', 'UInt32_', toUInt64(9999999), toUInt32(77)) on node Executing query select dictGetOrDefault('LocalClickHouse_hashed_', 'Int16_', toUInt64(9999999), toInt16(-66)) on local_node Executing query select dictGetUInt32OrDefault('MySQL_hashed_', 'UInt32_', toUInt64(9999999), toUInt32(77)) on node Executing query select dictGetInt16OrDefault('LocalClickHouse_hashed_', 'Int16_', toUInt64(9999999), toInt16(-66)) on local_node Executing query select dictGet('MySQL_hashed_', 'UInt64_', toUInt64(2)) on node Executing query select dictGet('LocalClickHouse_hashed_', 'Int32_', toUInt64(2)) on local_node Executing query select dictGetUInt64('MySQL_hashed_', 'UInt64_', toUInt64(2)) on node Executing query select dictGetInt32('LocalClickHouse_hashed_', 'Int32_', toUInt64(2)) on local_node Executing query select dictHas('MySQL_hashed_', toUInt64(2)) on node Executing query select dictHas('LocalClickHouse_hashed_', toUInt64(2)) on local_node Executing query select dictGetOrDefault('MySQL_hashed_', 'UInt64_', toUInt64(9999999), toUInt64(88)) on node Executing query select dictGetOrDefault('LocalClickHouse_hashed_', 'Int32_', toUInt64(9999999), toInt32(-77)) on local_node Executing query select dictGetUInt64OrDefault('MySQL_hashed_', 'UInt64_', toUInt64(9999999), toUInt64(88)) on node Executing query select dictGetInt32OrDefault('LocalClickHouse_hashed_', 'Int32_', toUInt64(9999999), toInt32(-77)) on local_node Executing query select dictGet('MySQL_hashed_', 'Int8_', toUInt64(2)) on node Executing query select dictGet('LocalClickHouse_hashed_', 'Int64_', toUInt64(2)) on local_node Executing query select dictGetInt8('MySQL_hashed_', 'Int8_', toUInt64(2)) on node Executing query select dictGetInt64('LocalClickHouse_hashed_', 'Int64_', toUInt64(2)) on local_node Executing query select dictHas('MySQL_hashed_', toUInt64(2)) on node Executing query select dictHas('LocalClickHouse_hashed_', toUInt64(2)) on local_node Executing query select dictGetOrDefault('MySQL_hashed_', 'Int8_', toUInt64(9999999), toInt8(-55)) on node Executing query select dictGetOrDefault('LocalClickHouse_hashed_', 'Int64_', toUInt64(9999999), toInt64(-88)) on local_node Executing query select dictGetInt8OrDefault('MySQL_hashed_', 'Int8_', toUInt64(9999999), toInt8(-55)) on node Executing query select dictGetInt64OrDefault('LocalClickHouse_hashed_', 'Int64_', toUInt64(9999999), toInt64(-88)) on local_node Executing query select dictGet('MySQL_hashed_', 'Int16_', toUInt64(2)) on node Executing query select dictGet('LocalClickHouse_hashed_', 'UUID_', toUInt64(2)) on local_node Executing query select dictGetInt16('MySQL_hashed_', 'Int16_', toUInt64(2)) on node Executing query select dictGetUUID('LocalClickHouse_hashed_', 'UUID_', toUInt64(2)) on local_node Executing query select dictHas('MySQL_hashed_', toUInt64(2)) on node Executing query select dictHas('LocalClickHouse_hashed_', toUInt64(2)) on local_node Executing query select dictGetOrDefault('MySQL_hashed_', 'Int16_', toUInt64(9999999), toInt16(-66)) on node Executing query select dictGetOrDefault('LocalClickHouse_hashed_', 'UUID_', toUInt64(9999999), toUUID('550e8400-0000-0000-0000-000000000000')) on local_node Executing query select dictGetInt16OrDefault('MySQL_hashed_', 'Int16_', toUInt64(9999999), toInt16(-66)) on node Executing query select dictGetUUIDOrDefault('LocalClickHouse_hashed_', 'UUID_', toUInt64(9999999), toUUID('550e8400-0000-0000-0000-000000000000')) on local_node Executing query select dictGet('MySQL_hashed_', 'Int32_', toUInt64(2)) on node Executing query select dictGet('LocalClickHouse_hashed_', 'Date_', toUInt64(2)) on local_node Executing query select dictGetInt32('MySQL_hashed_', 'Int32_', toUInt64(2)) on node Executing query select dictGetDate('LocalClickHouse_hashed_', 'Date_', toUInt64(2)) on local_node Executing query select dictHas('MySQL_hashed_', toUInt64(2)) on node Executing query select dictHas('LocalClickHouse_hashed_', toUInt64(2)) on local_node Executing query select dictGetOrDefault('MySQL_hashed_', 'Int32_', toUInt64(9999999), toInt32(-77)) on node Executing query select dictGetOrDefault('LocalClickHouse_hashed_', 'Date_', toUInt64(9999999), toDate('2018-12-30')) on local_node Executing query select dictGetInt32OrDefault('MySQL_hashed_', 'Int32_', toUInt64(9999999), toInt32(-77)) on node Executing query select dictGetDateOrDefault('LocalClickHouse_hashed_', 'Date_', toUInt64(9999999), toDate('2018-12-30')) on local_node Executing query select dictGet('MySQL_hashed_', 'Int64_', toUInt64(2)) on node Executing query select dictGet('LocalClickHouse_hashed_', 'DateTime_', toUInt64(2)) on local_node Executing query select dictGetInt64('MySQL_hashed_', 'Int64_', toUInt64(2)) on node Executing query select dictGetDateTime('LocalClickHouse_hashed_', 'DateTime_', toUInt64(2)) on local_node Executing query select dictHas('MySQL_hashed_', toUInt64(2)) on node Executing query select dictHas('LocalClickHouse_hashed_', toUInt64(2)) on local_node Executing query select dictGetOrDefault('MySQL_hashed_', 'Int64_', toUInt64(9999999), toInt64(-88)) on node Executing query select dictGetOrDefault('LocalClickHouse_hashed_', 'DateTime_', toUInt64(9999999), toDateTime('2018-12-30 00:00:00')) on local_node Executing query select dictGetDateTimeOrDefault('LocalClickHouse_hashed_', 'DateTime_', toUInt64(9999999), toDateTime('2018-12-30 00:00:00')) on local_node Executing query select dictGetInt64OrDefault('MySQL_hashed_', 'Int64_', toUInt64(9999999), toInt64(-88)) on node Executing query select dictGet('LocalClickHouse_hashed_', 'String_', toUInt64(2)) on local_node Executing query select dictGet('MySQL_hashed_', 'UUID_', toUInt64(2)) on node Executing query select dictGetUUID('MySQL_hashed_', 'UUID_', toUInt64(2)) on node Executing query select dictGetString('LocalClickHouse_hashed_', 'String_', toUInt64(2)) on local_node Executing query select dictHas('MySQL_hashed_', toUInt64(2)) on node Executing query select dictHas('LocalClickHouse_hashed_', toUInt64(2)) on local_node Executing query select dictGetOrDefault('MySQL_hashed_', 'UUID_', toUInt64(9999999), toUUID('550e8400-0000-0000-0000-000000000000')) on node Executing query select dictGetOrDefault('LocalClickHouse_hashed_', 'String_', toUInt64(9999999), toString('hi')) on local_node Executing query select dictGetUUIDOrDefault('MySQL_hashed_', 'UUID_', toUInt64(9999999), toUUID('550e8400-0000-0000-0000-000000000000')) on node Executing query select dictGetStringOrDefault('LocalClickHouse_hashed_', 'String_', toUInt64(9999999), toString('hi')) on local_node Executing query select dictGet('MySQL_hashed_', 'Date_', toUInt64(2)) on node Executing query select dictGet('LocalClickHouse_hashed_', 'Float32_', toUInt64(2)) on local_node Executing query select dictGetDate('MySQL_hashed_', 'Date_', toUInt64(2)) on node Executing query select dictGetFloat32('LocalClickHouse_hashed_', 'Float32_', toUInt64(2)) on local_node Executing query select dictHas('MySQL_hashed_', toUInt64(2)) on node Executing query select dictHas('LocalClickHouse_hashed_', toUInt64(2)) on local_node Executing query select dictGetOrDefault('MySQL_hashed_', 'Date_', toUInt64(9999999), toDate('2018-12-30')) on node Executing query select dictGetOrDefault('LocalClickHouse_hashed_', 'Float32_', toUInt64(9999999), toFloat32(555.11)) on local_node Executing query select dictGetDateOrDefault('MySQL_hashed_', 'Date_', toUInt64(9999999), toDate('2018-12-30')) on node Executing query select dictGetFloat32OrDefault('LocalClickHouse_hashed_', 'Float32_', toUInt64(9999999), toFloat32(555.11)) on local_node Executing query select dictGet('MySQL_hashed_', 'DateTime_', toUInt64(2)) on node Executing query select dictGet('LocalClickHouse_hashed_', 'Float64_', toUInt64(2)) on local_node Executing query select dictGetDateTime('MySQL_hashed_', 'DateTime_', toUInt64(2)) on node Executing query select dictGetFloat64('LocalClickHouse_hashed_', 'Float64_', toUInt64(2)) on local_node Executing query select dictHas('MySQL_hashed_', toUInt64(2)) on node Executing query select dictHas('LocalClickHouse_hashed_', toUInt64(2)) on local_node Executing query select dictGetOrDefault('MySQL_hashed_', 'DateTime_', toUInt64(9999999), toDateTime('2018-12-30 00:00:00')) on node Executing query select dictGetOrDefault('LocalClickHouse_hashed_', 'Float64_', toUInt64(9999999), toFloat64(777.11)) on local_node Executing query select dictGetDateTimeOrDefault('MySQL_hashed_', 'DateTime_', toUInt64(9999999), toDateTime('2018-12-30 00:00:00')) on node Executing query select dictGetFloat64OrDefault('LocalClickHouse_hashed_', 'Float64_', toUInt64(9999999), toFloat64(777.11)) on local_node Executing query select dictGet('MySQL_hashed_', 'String_', toUInt64(2)) on node Executing query select dictGet('LocalClickHouse_hashed_', 'ParentKeyField', toUInt64(2)) on local_node Executing query select dictGetString('MySQL_hashed_', 'String_', toUInt64(2)) on node Executing query select dictGetUInt64('LocalClickHouse_hashed_', 'ParentKeyField', toUInt64(2)) on local_node Executing query select dictHas('MySQL_hashed_', toUInt64(2)) on node Executing query select dictHas('LocalClickHouse_hashed_', toUInt64(2)) on local_node Executing query select dictGetOrDefault('MySQL_hashed_', 'String_', toUInt64(9999999), toString('hi')) on node Executing query select dictGetOrDefault('LocalClickHouse_hashed_', 'ParentKeyField', toUInt64(9999999), toUInt64(444)) on local_node Executing query select dictGetStringOrDefault('MySQL_hashed_', 'String_', toUInt64(9999999), toString('hi')) on node Executing query select dictGetUInt64OrDefault('LocalClickHouse_hashed_', 'ParentKeyField', toUInt64(9999999), toUInt64(444)) on local_node Executing query select dictGet('MySQL_hashed_', 'Float32_', toUInt64(2)) on node Executing query select dictGetHierarchy('LocalClickHouse_hashed_', toUInt64(1)) on local_node Executing query select dictGetFloat32('MySQL_hashed_', 'Float32_', toUInt64(2)) on node Executing query select dictGetHierarchy('LocalClickHouse_hashed_', toUInt64(2)) on local_node Executing query select dictHas('MySQL_hashed_', toUInt64(2)) on node Executing query select dictIsIn('LocalClickHouse_hashed_', toUInt64(1), toUInt64(2)) on local_node Executing query select dictGetOrDefault('MySQL_hashed_', 'Float32_', toUInt64(9999999), toFloat32(555.11)) on node Executing query select dictIsIn('LocalClickHouse_hashed_', toUInt64(2), toUInt64(1)) on local_node Executing query select dictGetFloat32OrDefault('MySQL_hashed_', 'Float32_', toUInt64(9999999), toFloat32(555.11)) on node Command:[docker compose --env-file /ClickHouse/tests/integration/test_dictionaries_all_layouts_separate_sources/_instances-clickhouse_local-0-gw5/.env --project-name roottestdictionariesalllayoutsseparatesourcesclickhouselocal-gw5 --file /ClickHouse/tests/integration/test_dictionaries_all_layouts_separate_sources/_instances-clickhouse_local-0-gw5/local_node/docker-compose.yml stop --timeout 20] [gw5] PASSED test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_simple[hashed] Executing query select dictGet('MySQL_hashed_', 'Float64_', toUInt64(2)) on node Executing query select dictGetFloat64('MySQL_hashed_', 'Float64_', toUInt64(2)) on node Executing query select dictHas('MySQL_hashed_', toUInt64(2)) on node Executing query select dictGetOrDefault('MySQL_hashed_', 'Float64_', toUInt64(9999999), toFloat64(777.11)) on node Executing query select dictGetFloat64OrDefault('MySQL_hashed_', 'Float64_', toUInt64(9999999), toFloat64(777.11)) on node Executing query select dictGet('MySQL_hashed_', 'ParentKeyField', toUInt64(2)) on node Executing query select dictGetUInt64('MySQL_hashed_', 'ParentKeyField', toUInt64(2)) on node Executing query select dictHas('MySQL_hashed_', toUInt64(2)) on node Executing query select dictGetOrDefault('MySQL_hashed_', 'ParentKeyField', toUInt64(9999999), toUInt64(444)) on node Executing query select dictGetUInt64OrDefault('MySQL_hashed_', 'ParentKeyField', toUInt64(9999999), toUInt64(444)) on node Executing query select dictGetHierarchy('MySQL_hashed_', toUInt64(1)) on node Executing query select dictGetHierarchy('MySQL_hashed_', toUInt64(2)) on node Executing query select dictIsIn('MySQL_hashed_', toUInt64(1), toUInt64(2)) on node Executing query select dictIsIn('MySQL_hashed_', toUInt64(2), toUInt64(1)) on node Command:[docker compose --env-file /ClickHouse/tests/integration/test_dictionaries_all_layouts_separate_sources/_instances-mysql-0-gw3/.env --project-name roottestdictionariesalllayoutsseparatesourcesmysql-gw3 --file /ClickHouse/tests/integration/test_dictionaries_all_layouts_separate_sources/_instances-mysql-0-gw3/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_mysql_8_0.yml stop --timeout 20] [gw3] PASSED test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_simple[hashed] Stderr: Container roottestdictionariesalllayoutsseparatesourcesclickhouselocal-gw5-local_node-1 Stopping Stderr: Container roottestdictionariesalllayoutsseparatesourcesclickhouselocal-gw5-local_node-1 Stopped Command:[bash -c [ -f /ClickHouse/tests/integration/test_dictionaries_all_layouts_separate_sources/_instances-clickhouse_local-0-gw5/local_node/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_dictionaries_all_layouts_separate_sources/_instances-clickhouse_local-0-gw5/local_node/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[docker compose --env-file /ClickHouse/tests/integration/test_dictionaries_all_layouts_separate_sources/_instances-clickhouse_local-0-gw5/.env --project-name roottestdictionariesalllayoutsseparatesourcesclickhouselocal-gw5 --file /ClickHouse/tests/integration/test_dictionaries_all_layouts_separate_sources/_instances-clickhouse_local-0-gw5/local_node/docker-compose.yml down --volumes] Stderr: Container roottestdictionariesalllayoutsseparatesourcesclickhouselocal-gw5-local_node-1 Stopping Stderr: Container roottestdictionariesalllayoutsseparatesourcesclickhouselocal-gw5-local_node-1 Stopped Stderr: Container roottestdictionariesalllayoutsseparatesourcesclickhouselocal-gw5-local_node-1 Removing Stderr: Container roottestdictionariesalllayoutsseparatesourcesclickhouselocal-gw5-local_node-1 Removed Stderr: Network roottestdictionariesalllayoutsseparatesourcesclickhouselocal-gw5_default Removing Stderr: Network roottestdictionariesalllayoutsseparatesourcesclickhouselocal-gw5_default Removed Cleanup called Docker networks for project roottestdictionariesalllayoutsseparatesourcesclickhouselocal-gw5 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestdictionariesalllayoutsseparatesourcesclickhouselocal-gw5 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestdictionariesalllayoutsseparatesourcesclickhouselocal-gw5 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestdictionariesalllayoutsseparatesourcesclickhouselocal-gw5-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} No running containers for project: roottestdictionariesalllayoutsseparatesourcesclickhouselocal-gw5 Trying to prune unused networks... Trying to prune unused images... Command:[docker image prune -f] Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Stdout:2 Command:[docker volume prune -f] Stdout:Total reclaimed space: 0B Volumes pruned: 2 Stderr: Container roottestdictionariesalllayoutsseparatesourcesmysql-gw3-node-1 Stopping Stderr: Container roottestdictionariesalllayoutsseparatesourcesmysql-gw3-node-1 Stopped Stderr: Container roottestdictionariesalllayoutsseparatesourcesmysql-gw3-mysql80-1 Stopping Stderr: Container roottestdictionariesalllayoutsseparatesourcesmysql-gw3-mysql80-1 Stopped Command:[bash -c [ -f /ClickHouse/tests/integration/test_dictionaries_all_layouts_separate_sources/_instances-mysql-0-gw3/node/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_dictionaries_all_layouts_separate_sources/_instances-mysql-0-gw3/node/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[docker compose --env-file /ClickHouse/tests/integration/test_dictionaries_all_layouts_separate_sources/_instances-mysql-0-gw3/.env --project-name roottestdictionariesalllayoutsseparatesourcesmysql-gw3 --file /ClickHouse/tests/integration/test_dictionaries_all_layouts_separate_sources/_instances-mysql-0-gw3/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_mysql_8_0.yml down --volumes] Stderr: Container roottestdictionariesalllayoutsseparatesourcesmysql-gw3-node-1 Stopping Stderr: Container roottestdictionariesalllayoutsseparatesourcesmysql-gw3-node-1 Stopped Stderr: Container roottestdictionariesalllayoutsseparatesourcesmysql-gw3-node-1 Removing Stderr: Container roottestdictionariesalllayoutsseparatesourcesmysql-gw3-node-1 Removed Stderr: Container roottestdictionariesalllayoutsseparatesourcesmysql-gw3-mysql80-1 Stopping Stderr: Container roottestdictionariesalllayoutsseparatesourcesmysql-gw3-mysql80-1 Stopped Stderr: Container roottestdictionariesalllayoutsseparatesourcesmysql-gw3-mysql80-1 Removing Stderr: Container roottestdictionariesalllayoutsseparatesourcesmysql-gw3-mysql80-1 Removed Stderr: Network roottestdictionariesalllayoutsseparatesourcesmysql-gw3_default Removing Stderr: Network roottestdictionariesalllayoutsseparatesourcesmysql-gw3_default Removed Cleanup called Docker networks for project roottestdictionariesalllayoutsseparatesourcesmysql-gw3 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestdictionariesalllayoutsseparatesourcesmysql-gw3 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestdictionariesalllayoutsseparatesourcesmysql-gw3 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestdictionariesalllayoutsseparatesourcesmysql-gw3-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} No running containers for project: roottestdictionariesalllayoutsseparatesourcesmysql-gw3 Trying to prune unused networks... Trying to prune unused images... Command:[docker image prune -f] Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Stdout:1 Volumes pruned: 1 ============================== slowest durations =============================== 176.42s setup test_azure_blob_storage_zero_copy_replication/test.py::test_zero_copy_replication 139.61s call test_async_load_databases/test.py::test_async_load_system_database 49.17s call test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_simple[direct] 44.96s call test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_complex[complex_key_direct] 44.75s call test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_simple[direct] 44.38s call test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_simple[cache] 43.98s call test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_simple[cache] 42.86s call test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_simple[hashed] 42.61s call test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_complex[complex_key_cache] 42.32s call test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_simple[hashed] 42.27s call test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_simple[flat] 42.11s call test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_simple[flat] 41.75s call test_async_load_databases/test.py::test_dependent_tables 41.32s call test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_complex[complex_key_direct] 41.30s call test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_complex[complex_key_cache] 40.56s call test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_complex[complex_key_hashed] 40.54s call test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_complex[complex_key_hashed] 35.84s setup test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_complex[complex_key_cache] 35.13s call test_async_load_databases/test.py::test_dict_get_data 27.36s call test_async_load_databases/test.py::test_multiple_tables 27.23s setup test_dictionaries_redis/test.py::test_redis_dictionaries[DateTime_] 26.42s call test_cluster_all_replicas/test.py::test_skip_unavailable_replica[two_shards_three_nodes] 25.20s setup test_cluster_all_replicas/test.py::test_cluster 23.88s call test_dictionaries_redis/test.py::test_redis_dictionaries[DateTime_] 23.37s setup test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[http-False] 23.22s setup test_config_substitutions/test.py::test_allow_databases 22.68s call test_dictionaries_redis/test.py::test_redis_dictionaries[Int32_] 22.43s call test_dictionaries_redis/test.py::test_redis_dictionaries[Float64_] 22.37s call test_dictionaries_redis/test.py::test_redis_dictionaries[String_] 22.36s call test_dictionaries_redis/test.py::test_redis_dictionaries[Date_] 22.24s call test_dictionaries_redis/test.py::test_redis_dictionaries[Float32_] 22.04s call test_dictionaries_redis/test.py::test_redis_dictionaries[Int16_] 21.98s teardown test_database_backup/test.py::test_database_backup_table[File('test_table_backup_file')] 21.87s call test_dictionaries_redis/test.py::test_redis_dictionaries[Int8_] 21.68s call test_dictionaries_redis/test.py::test_redis_dictionaries[Int64_] 21.24s setup test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_complex[complex_key_cache] 20.23s setup test_analyzer_compatibility/test.py::test_two_new_versions 19.79s teardown test_azure_blob_storage_zero_copy_replication/test.py::test_zero_copy_replication 18.83s call test_backup_restore_on_cluster/test.py::test_shutdown_waits_for_backup 18.09s setup test_database_backup/test.py::test_database_backup_database[Disk('backup_disk_local', 'test_database_backup')] 17.50s call test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_ranged[range_hashed] 16.92s setup test_backup_restore/test.py::test_attach_partition 16.90s call test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_ranged[range_hashed] 15.41s teardown test_config_substitutions/test.py::test_include_config 15.38s setup test_aggregation_memory_efficient/test.py::test_remote 15.00s setup test_async_load_databases/test.py::test_async_load_system_database 14.78s teardown test_cluster_all_replicas/test.py::test_skip_unavailable_replica[two_shards_three_nodes] 14.44s setup test_allow_feature_tier/test.py::test_allow_feature_tier_in_general_settings 14.25s setup test_attach_table_normalizer/test.py::test_attach_substr 14.00s call test_allow_feature_tier/test.py::test_allow_feature_tier_in_user 13.60s setup test_config_corresponding_root/test.py::test_work 13.38s setup test_composable_protocols/test.py::test_connections 13.34s call test_allow_feature_tier/test.py::test_allow_feature_tier_in_mergetree_settings 12.29s call test_database_backup/test.py::test_database_backup_database[Disk('backup_disk_s3_plain', 'test_database_backup')] 12.17s call test_database_backup/test.py::test_database_backup_database[Disk('backup_disk_local', 'test_database_backup')] 11.19s setup test_config_hide_in_preprocessed/test.py::test_hide_in_preprocessed 10.66s call test_database_backup/test.py::test_database_backup_table[Disk('backup_disk_s3_plain', 'test_table_backup')] 10.18s call test_database_backup/test.py::test_database_backup_database[Disk('backup_disk_object_storage_local_plain', 'test_database_backup')] 10.10s call test_database_backup/test.py::test_database_backup_table[File('test_table_backup_file')] 9.81s call test_database_backup/test.py::test_database_backup_table[Disk('backup_disk_object_storage_local_plain', 'test_table_backup')] 9.73s call test_database_backup/test.py::test_database_backup_table[Disk('backup_disk_local', 'test_table_backup')] 9.70s call test_database_backup/test.py::test_database_backup_database[File('test_database_backup_file')] 9.30s call test_config_substitutions/test.py::test_config_multiple_zk_substitutions 9.11s call test_backup_restore_on_cluster/test.py::test_replicated_database_async 8.83s call test_backup_restore/test.py::test_attach_partition 8.63s teardown test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_simple[hashed] 8.62s teardown test_backup_restore_on_cluster/test.py::test_tables_dependency 8.01s call test_backup_restore_on_cluster/test.py::test_tables_dependency 7.90s call test_backup_restore/test.py::test_replace_partition 7.47s call test_analyzer_compatibility/test.py::test_two_new_versions 7.43s call test_backup_restore_on_cluster/test.py::test_required_privileges 7.03s call test_backup_restore_on_cluster/test.py::test_keeper_value_max_size 6.71s teardown test_analyzer_compatibility/test.py::test_two_new_versions 6.52s call test_cluster_all_replicas/test.py::test_error_on_unavailable_replica[one_shard_three_nodes] 6.45s call test_backup_restore_on_cluster/test.py::test_replicated_database_compare_parts 6.34s call test_backup_restore/test.py::test_restore 6.32s call test_cluster_all_replicas/test.py::test_error_on_unavailable_replica[two_shards_three_nodes] 6.27s call test_cluster_all_replicas/test.py::test_skip_unavailable_replica[one_shard_three_nodes] 6.25s call test_backup_restore_on_cluster/test.py::test_mutation 6.22s call test_backup_restore_on_cluster/test.py::test_replicated_database 6.17s setup test_config_decryption/test.py::test_successful_decryption_xml 5.92s call test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[http-True] 5.47s call test_backup_restore_on_cluster/test.py::test_replicated_database_with_special_macro_in_zk_path[uuid] 5.45s call test_backup_restore_on_cluster/test.py::test_backup_restore_on_single_replica 5.44s teardown test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_simple[hashed] 5.33s call test_backup_restore_on_cluster/test.py::test_table_in_replicated_database_with_not_synced_def 5.29s call test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_def 5.21s call test_backup_restore_on_cluster/test.py::test_replicated_database_with_special_macro_in_zk_path[database] 4.94s teardown test_dictionaries_redis/test.py::test_redis_dictionaries[String_] 4.76s call test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_insert 4.70s call test_backup_restore_on_cluster/test.py::test_replicated_table 4.64s call test_backup_restore_on_cluster/test.py::test_replicated_table_after_alters 4.64s call test_backup_restore_on_cluster/test.py::test_system_functions 4.60s call test_attach_table_normalizer/test.py::test_attach_substr_restart 4.56s call test_backup_restore_on_cluster/test.py::test_file_deduplication 4.50s call test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_merge 4.49s call test_backup_restore_on_cluster/test.py::test_system_users 4.34s call test_backup_restore_on_cluster/test.py::test_replicated_table_with_uuid_in_zkpath 4.18s call test_backup_restore_on_cluster/test.py::test_projection 4.13s call test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[native-True] 4.11s call test_backup_restore_on_cluster/test.py::test_replicated_table_restored_into_bigger_cluster 4.10s call test_aggregation_memory_efficient/test.py::test_remote 4.06s teardown test_config_decryption/test.py::test_successful_decryption_yaml 4.00s call test_backup_restore_on_cluster/test.py::test_table_with_parts_in_queue_considered_non_empty 3.99s call test_config_substitutions/test.py::test_allow_databases 3.92s teardown test_backup_restore/test.py::test_restore 3.91s call test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[http-False] 3.76s call test_allow_feature_tier/test.py::test_allow_feature_tier_in_general_settings 3.46s call test_azure_blob_storage_zero_copy_replication/test.py::test_zero_copy_replication 3.11s call test_backup_restore_on_cluster/test.py::test_replicated_table_restored_into_smaller_cluster 3.08s teardown test_aggregation_memory_efficient/test.py::test_remote 3.06s teardown test_config_hide_in_preprocessed/test.py::test_hide_in_preprocessed 3.04s call test_backup_restore_on_cluster/test.py::test_different_tables_on_nodes 2.98s call test_backup_restore_on_cluster/test.py::test_empty_replicated_table 2.80s call test_config_substitutions/test.py::test_include_config 2.69s teardown test_composable_protocols/test.py::test_proxy_1 2.59s call test_config_substitutions/test.py::test_config 2.49s call test_allow_feature_tier/test.py::test_allow_feature_tier_in_mergetree_settings_with_old_compatibility 2.18s call test_composable_protocols/test.py::test_proxy_1 2.08s teardown test_backup_restore_on_cluster/test.py::test_keeper_value_max_size 2.05s teardown test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[http-False] 2.03s teardown test_backup_restore_on_cluster/test.py::test_replicated_database_async 2.03s teardown test_backup_restore_on_cluster/test.py::test_replicated_table_with_uuid_in_zkpath 1.98s teardown test_backup_restore_on_cluster/test.py::test_replicated_table_restored_into_smaller_cluster 1.94s call test_allow_feature_tier/test.py::test_it_is_possible_to_enable_experimental_settings_in_default_profile 1.94s teardown test_backup_restore_on_cluster/test.py::test_table_in_replicated_database_with_not_synced_def 1.93s teardown test_backup_restore_on_cluster/test.py::test_shutdown_waits_for_backup 1.93s teardown test_backup_restore_on_cluster/test.py::test_replicated_table_restored_into_bigger_cluster 1.93s teardown test_backup_restore_on_cluster/test.py::test_replicated_database_with_special_macro_in_zk_path[database] 1.89s teardown test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[http-True] 1.89s teardown test_backup_restore_on_cluster/test.py::test_replicated_database 1.89s teardown test_backup_restore_on_cluster/test.py::test_replicated_database_with_special_macro_in_zk_path[uuid] 1.88s teardown test_backup_restore_on_cluster/test.py::test_mutation 1.88s teardown test_backup_restore_on_cluster/test.py::test_file_deduplication 1.88s teardown test_backup_restore_on_cluster/test.py::test_different_tables_on_nodes 1.88s teardown test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[native-True] 1.88s teardown test_backup_restore_on_cluster/test.py::test_replicated_table_after_alters 1.88s teardown test_backup_restore_on_cluster/test.py::test_table_with_parts_in_queue_considered_non_empty 1.88s teardown test_backup_restore_on_cluster/test.py::test_backup_restore_on_single_replica 1.88s teardown test_backup_restore_on_cluster/test.py::test_projection 1.88s teardown test_backup_restore_on_cluster/test.py::test_required_privileges 1.88s teardown test_backup_restore_on_cluster/test.py::test_replicated_database_compare_parts 1.88s teardown test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_def 1.88s teardown test_backup_restore_on_cluster/test.py::test_replicated_table 1.88s teardown test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_merge 1.84s teardown test_backup_restore_on_cluster/test.py::test_empty_replicated_table 1.83s teardown test_backup_restore_on_cluster/test.py::test_system_users 1.83s teardown test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_insert 1.83s teardown test_backup_restore_on_cluster/test.py::test_get_error_from_other_host 1.83s teardown test_backup_restore_on_cluster/test.py::test_system_functions 1.74s call test_config_substitutions/test.py::test_config_from_env_overrides 1.41s call test_config_hide_in_preprocessed/test.py::test_hide_in_preprocessed 1.34s teardown test_allow_feature_tier/test.py::test_it_is_possible_to_enable_experimental_settings_in_default_profile 1.31s teardown test_attach_table_normalizer/test.py::test_attach_substr_restart 1.25s call test_backup_restore_on_cluster/test.py::test_get_error_from_other_host 1.24s teardown test_async_load_databases/test.py::test_multiple_tables 1.17s call test_cluster_all_replicas/test.py::test_cluster 1.11s call test_composable_protocols/test.py::test_connections 1.09s call test_config_substitutions/test.py::test_config_merge_from_env_overrides 0.92s call test_attach_table_normalizer/test.py::test_attach_substr 0.85s call test_cluster_all_replicas/test.py::test_global_in 0.59s call test_config_decryption/test.py::test_successful_decryption_xml 0.58s call test_config_decryption/test.py::test_successful_decryption_yaml 0.01s call test_composable_protocols/test.py::test_http_proxy_1 0.00s setup test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_def 0.00s setup test_backup_restore_on_cluster/test.py::test_mutation 0.00s setup test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_simple[hashed] 0.00s teardown test_async_load_databases/test.py::test_async_load_system_database 0.00s setup test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_complex[complex_key_hashed] 0.00s setup test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_merge 0.00s teardown test_dictionaries_redis/test.py::test_redis_dictionaries[DateTime_] 0.00s teardown test_config_substitutions/test.py::test_allow_databases 0.00s setup test_backup_restore_on_cluster/test.py::test_replicated_database_with_special_macro_in_zk_path[uuid] 0.00s setup test_backup_restore_on_cluster/test.py::test_projection 0.00s setup test_dictionaries_redis/test.py::test_redis_dictionaries[Date_] 0.00s setup test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_ranged[range_hashed] 0.00s setup test_backup_restore_on_cluster/test.py::test_replicated_table_restored_into_bigger_cluster 0.00s setup test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[http-True] 0.00s teardown test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_complex[complex_key_direct] 0.00s teardown test_config_corresponding_root/test.py::test_work 0.00s setup test_allow_feature_tier/test.py::test_allow_feature_tier_in_user 0.00s setup test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_complex[complex_key_direct] 0.00s setup test_backup_restore_on_cluster/test.py::test_replicated_database_async 0.00s setup test_backup_restore_on_cluster/test.py::test_replicated_database_with_special_macro_in_zk_path[database] 0.00s teardown test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_simple[flat] 0.00s setup test_dictionaries_redis/test.py::test_redis_dictionaries[Int64_] 0.00s setup test_backup_restore_on_cluster/test.py::test_shutdown_waits_for_backup 0.00s setup test_backup_restore_on_cluster/test.py::test_table_with_parts_in_queue_considered_non_empty 0.00s setup test_database_backup/test.py::test_database_backup_table[Disk('backup_disk_local', 'test_table_backup')] 0.00s setup test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_simple[direct] 0.00s setup test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_complex[complex_key_direct] 0.00s teardown test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_complex[complex_key_cache] 0.00s setup test_backup_restore_on_cluster/test.py::test_get_error_from_other_host 0.00s setup test_backup_restore_on_cluster/test.py::test_different_tables_on_nodes 0.00s teardown test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_complex[complex_key_hashed] 0.00s setup test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_insert 0.00s setup test_backup_restore_on_cluster/test.py::test_system_functions 0.00s setup test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[native-True] 0.00s setup test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_simple[hashed] 0.00s setup test_backup_restore_on_cluster/test.py::test_table_in_replicated_database_with_not_synced_def 0.00s setup test_backup_restore_on_cluster/test.py::test_replicated_table_restored_into_smaller_cluster 0.00s setup test_async_load_databases/test.py::test_dependent_tables 0.00s setup test_dictionaries_redis/test.py::test_redis_dictionaries[Float64_] 0.00s setup test_backup_restore_on_cluster/test.py::test_tables_dependency 0.00s setup test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_complex[complex_key_hashed] 0.00s setup test_database_backup/test.py::test_database_backup_table[Disk('backup_disk_object_storage_local_plain', 'test_table_backup')] 0.00s setup test_backup_restore_on_cluster/test.py::test_replicated_table 0.00s setup test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_simple[flat] 0.00s setup test_database_backup/test.py::test_database_backup_database[Disk('backup_disk_s3_plain', 'test_database_backup')] 0.00s setup test_dictionaries_redis/test.py::test_redis_dictionaries[Int32_] 0.00s setup test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_simple[direct] 0.00s teardown test_allow_feature_tier/test.py::test_allow_feature_tier_in_general_settings 0.00s setup test_backup_restore_on_cluster/test.py::test_keeper_value_max_size 0.00s setup test_database_backup/test.py::test_database_backup_table[Disk('backup_disk_s3_plain', 'test_table_backup')] 0.00s setup test_config_substitutions/test.py::test_config_from_env_overrides 0.00s setup test_backup_restore_on_cluster/test.py::test_backup_restore_on_single_replica 0.00s setup test_backup_restore_on_cluster/test.py::test_replicated_table_with_uuid_in_zkpath 0.00s teardown test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_complex[complex_key_cache] 0.00s setup test_backup_restore_on_cluster/test.py::test_replicated_table_after_alters 0.00s setup test_dictionaries_redis/test.py::test_redis_dictionaries[Int16_] 0.00s setup test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_simple[cache] 0.00s call test_config_corresponding_root/test.py::test_work 0.00s setup test_composable_protocols/test.py::test_proxy_1 0.00s setup test_backup_restore_on_cluster/test.py::test_replicated_database 0.00s setup test_database_backup/test.py::test_database_backup_database[Disk('backup_disk_object_storage_local_plain', 'test_database_backup')] 0.00s teardown test_allow_feature_tier/test.py::test_allow_feature_tier_in_user 0.00s setup test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_simple[flat] 0.00s teardown test_dictionaries_redis/test.py::test_redis_dictionaries[Int8_] 0.00s teardown test_database_backup/test.py::test_database_backup_table[Disk('backup_disk_local', 'test_table_backup')] 0.00s setup test_dictionaries_redis/test.py::test_redis_dictionaries[String_] 0.00s teardown test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_simple[cache] 0.00s teardown test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_simple[cache] 0.00s setup test_composable_protocols/test.py::test_http_proxy_1 0.00s setup test_dictionaries_redis/test.py::test_redis_dictionaries[Int8_] 0.00s setup test_cluster_all_replicas/test.py::test_skip_unavailable_replica[two_shards_three_nodes] 0.00s setup test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_ranged[range_hashed] 0.00s setup test_backup_restore_on_cluster/test.py::test_file_deduplication 0.00s setup test_dictionaries_redis/test.py::test_redis_dictionaries[Float32_] 0.00s setup test_allow_feature_tier/test.py::test_allow_feature_tier_in_mergetree_settings 0.00s setup test_allow_feature_tier/test.py::test_it_is_possible_to_enable_experimental_settings_in_default_profile 0.00s setup test_backup_restore_on_cluster/test.py::test_replicated_database_compare_parts 0.00s teardown test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_complex[complex_key_direct] 0.00s teardown test_dictionaries_redis/test.py::test_redis_dictionaries[Float64_] 0.00s teardown test_dictionaries_redis/test.py::test_redis_dictionaries[Int16_] 0.00s teardown test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_simple[direct] 0.00s setup test_backup_restore_on_cluster/test.py::test_system_users 0.00s teardown test_async_load_databases/test.py::test_dependent_tables 0.00s teardown test_composable_protocols/test.py::test_connections 0.00s setup test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_simple[cache] 0.00s setup test_backup_restore_on_cluster/test.py::test_required_privileges 0.00s teardown test_config_substitutions/test.py::test_config_multiple_zk_substitutions 0.00s setup test_database_backup/test.py::test_database_backup_table[File('test_table_backup_file')] 0.00s teardown test_database_backup/test.py::test_database_backup_database[Disk('backup_disk_local', 'test_database_backup')] 0.00s setup test_cluster_all_replicas/test.py::test_skip_unavailable_replica[one_shard_three_nodes] 0.00s teardown test_database_backup/test.py::test_database_backup_database[File('test_database_backup_file')] 0.00s teardown test_database_backup/test.py::test_database_backup_table[Disk('backup_disk_object_storage_local_plain', 'test_table_backup')] 0.00s teardown test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_simple[direct] 0.00s setup test_config_substitutions/test.py::test_include_config 0.00s setup test_database_backup/test.py::test_database_backup_database[File('test_database_backup_file')] 0.00s teardown test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_complex[complex_key_hashed] 0.00s setup test_backup_restore_on_cluster/test.py::test_empty_replicated_table 0.00s setup test_async_load_databases/test.py::test_dict_get_data 0.00s setup test_cluster_all_replicas/test.py::test_error_on_unavailable_replica[two_shards_three_nodes] 0.00s setup test_cluster_all_replicas/test.py::test_error_on_unavailable_replica[one_shard_three_nodes] 0.00s setup test_backup_restore/test.py::test_replace_partition 0.00s teardown test_dictionaries_redis/test.py::test_redis_dictionaries[Float32_] 0.00s teardown test_dictionaries_redis/test.py::test_redis_dictionaries[Int64_] 0.00s teardown test_config_substitutions/test.py::test_config 0.00s teardown test_database_backup/test.py::test_database_backup_table[Disk('backup_disk_s3_plain', 'test_table_backup')] 0.00s teardown test_cluster_all_replicas/test.py::test_cluster 0.00s teardown test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_ranged[range_hashed] 0.00s teardown test_dictionaries_redis/test.py::test_redis_dictionaries[Int32_] 0.00s teardown test_backup_restore/test.py::test_attach_partition 0.00s teardown test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_ranged[range_hashed] 0.00s teardown test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_simple[flat] 0.00s teardown test_composable_protocols/test.py::test_http_proxy_1 0.00s setup test_backup_restore/test.py::test_restore 0.00s setup test_attach_table_normalizer/test.py::test_attach_substr_restart 0.00s teardown test_attach_table_normalizer/test.py::test_attach_substr 0.00s teardown test_config_decryption/test.py::test_successful_decryption_xml 0.00s setup test_async_load_databases/test.py::test_multiple_tables 0.00s setup test_config_substitutions/test.py::test_config 0.00s teardown test_dictionaries_redis/test.py::test_redis_dictionaries[Date_] 0.00s teardown test_allow_feature_tier/test.py::test_allow_feature_tier_in_mergetree_settings_with_old_compatibility 0.00s setup test_config_substitutions/test.py::test_config_merge_from_env_overrides 0.00s setup test_allow_feature_tier/test.py::test_allow_feature_tier_in_mergetree_settings_with_old_compatibility 0.00s teardown test_database_backup/test.py::test_database_backup_database[Disk('backup_disk_object_storage_local_plain', 'test_database_backup')] 0.00s teardown test_cluster_all_replicas/test.py::test_skip_unavailable_replica[one_shard_three_nodes] 0.00s teardown test_database_backup/test.py::test_database_backup_database[Disk('backup_disk_s3_plain', 'test_database_backup')] 0.00s setup test_cluster_all_replicas/test.py::test_global_in 0.00s teardown test_async_load_databases/test.py::test_dict_get_data 0.00s setup test_config_decryption/test.py::test_successful_decryption_yaml 0.00s setup test_config_substitutions/test.py::test_config_multiple_zk_substitutions 0.00s teardown test_cluster_all_replicas/test.py::test_error_on_unavailable_replica[one_shard_three_nodes] 0.00s teardown test_backup_restore/test.py::test_replace_partition 0.00s teardown test_allow_feature_tier/test.py::test_allow_feature_tier_in_mergetree_settings 0.00s teardown test_cluster_all_replicas/test.py::test_global_in 0.00s teardown test_cluster_all_replicas/test.py::test_error_on_unavailable_replica[two_shards_three_nodes] 0.00s teardown test_config_substitutions/test.py::test_config_from_env_overrides 0.00s teardown test_config_substitutions/test.py::test_config_merge_from_env_overrides =========================== short test summary info ============================ PASSED test_allow_feature_tier/test.py::test_allow_feature_tier_in_general_settings PASSED test_backup_restore/test.py::test_attach_partition PASSED test_cluster_all_replicas/test.py::test_cluster PASSED test_config_substitutions/test.py::test_allow_databases PASSED test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[http-False] PASSED test_config_substitutions/test.py::test_config PASSED test_database_backup/test.py::test_database_backup_database[Disk('backup_disk_local', 'test_database_backup')] PASSED test_allow_feature_tier/test.py::test_allow_feature_tier_in_mergetree_settings PASSED test_config_substitutions/test.py::test_config_from_env_overrides PASSED test_config_substitutions/test.py::test_config_merge_from_env_overrides PASSED test_cluster_all_replicas/test.py::test_error_on_unavailable_replica[one_shard_three_nodes] PASSED test_backup_restore/test.py::test_replace_partition PASSED test_allow_feature_tier/test.py::test_allow_feature_tier_in_mergetree_settings_with_old_compatibility PASSED test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[http-True] PASSED test_cluster_all_replicas/test.py::test_error_on_unavailable_replica[two_shards_three_nodes] PASSED test_backup_restore/test.py::test_restore PASSED test_cluster_all_replicas/test.py::test_global_in PASSED test_database_backup/test.py::test_database_backup_database[Disk('backup_disk_object_storage_local_plain', 'test_database_backup')] PASSED test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[native-True] PASSED test_config_substitutions/test.py::test_config_multiple_zk_substitutions PASSED test_config_substitutions/test.py::test_include_config PASSED test_cluster_all_replicas/test.py::test_skip_unavailable_replica[one_shard_three_nodes] PASSED test_allow_feature_tier/test.py::test_allow_feature_tier_in_user PASSED test_backup_restore_on_cluster/test.py::test_backup_restore_on_single_replica PASSED test_allow_feature_tier/test.py::test_it_is_possible_to_enable_experimental_settings_in_default_profile PASSED test_dictionaries_redis/test.py::test_redis_dictionaries[DateTime_] PASSED test_database_backup/test.py::test_database_backup_database[Disk('backup_disk_s3_plain', 'test_database_backup')] PASSED test_backup_restore_on_cluster/test.py::test_different_tables_on_nodes PASSED test_config_decryption/test.py::test_successful_decryption_xml PASSED test_backup_restore_on_cluster/test.py::test_empty_replicated_table PASSED test_composable_protocols/test.py::test_connections PASSED test_composable_protocols/test.py::test_http_proxy_1 PASSED test_config_decryption/test.py::test_successful_decryption_yaml PASSED test_composable_protocols/test.py::test_proxy_1 PASSED test_database_backup/test.py::test_database_backup_database[File('test_database_backup_file')] PASSED test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_complex[complex_key_cache] PASSED test_backup_restore_on_cluster/test.py::test_file_deduplication PASSED test_backup_restore_on_cluster/test.py::test_get_error_from_other_host PASSED test_database_backup/test.py::test_database_backup_table[Disk('backup_disk_local', 'test_table_backup')] PASSED test_cluster_all_replicas/test.py::test_skip_unavailable_replica[two_shards_three_nodes] PASSED test_dictionaries_redis/test.py::test_redis_dictionaries[Date_] PASSED test_attach_table_normalizer/test.py::test_attach_substr PASSED test_config_hide_in_preprocessed/test.py::test_hide_in_preprocessed PASSED test_backup_restore_on_cluster/test.py::test_keeper_value_max_size PASSED test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_complex[complex_key_cache] PASSED test_attach_table_normalizer/test.py::test_attach_substr_restart PASSED test_database_backup/test.py::test_database_backup_table[Disk('backup_disk_object_storage_local_plain', 'test_table_backup')] PASSED test_backup_restore_on_cluster/test.py::test_mutation PASSED test_backup_restore_on_cluster/test.py::test_projection PASSED test_database_backup/test.py::test_database_backup_table[Disk('backup_disk_s3_plain', 'test_table_backup')] PASSED test_dictionaries_redis/test.py::test_redis_dictionaries[Float32_] PASSED test_backup_restore_on_cluster/test.py::test_replicated_database PASSED test_database_backup/test.py::test_database_backup_table[File('test_table_backup_file')] PASSED test_aggregation_memory_efficient/test.py::test_remote PASSED test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_complex[complex_key_direct] PASSED test_backup_restore_on_cluster/test.py::test_replicated_database_async PASSED test_dictionaries_redis/test.py::test_redis_dictionaries[Float64_] PASSED test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_complex[complex_key_direct] PASSED test_backup_restore_on_cluster/test.py::test_replicated_database_compare_parts PASSED test_backup_restore_on_cluster/test.py::test_replicated_database_with_special_macro_in_zk_path[database] PASSED test_backup_restore_on_cluster/test.py::test_replicated_database_with_special_macro_in_zk_path[uuid] PASSED test_analyzer_compatibility/test.py::test_two_new_versions PASSED test_backup_restore_on_cluster/test.py::test_replicated_table PASSED test_dictionaries_redis/test.py::test_redis_dictionaries[Int16_] PASSED test_backup_restore_on_cluster/test.py::test_replicated_table_after_alters PASSED test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_complex[complex_key_hashed] PASSED test_backup_restore_on_cluster/test.py::test_replicated_table_restored_into_bigger_cluster PASSED test_async_load_databases/test.py::test_async_load_system_database PASSED test_backup_restore_on_cluster/test.py::test_replicated_table_restored_into_smaller_cluster PASSED test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_complex[complex_key_hashed] PASSED test_dictionaries_redis/test.py::test_redis_dictionaries[Int32_] PASSED test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_def PASSED test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_ranged[range_hashed] PASSED test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_insert PASSED test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_ranged[range_hashed] PASSED test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_merge PASSED test_backup_restore_on_cluster/test.py::test_replicated_table_with_uuid_in_zkpath PASSED test_dictionaries_redis/test.py::test_redis_dictionaries[Int64_] PASSED test_backup_restore_on_cluster/test.py::test_required_privileges PASSED test_async_load_databases/test.py::test_dependent_tables PASSED test_dictionaries_redis/test.py::test_redis_dictionaries[Int8_] PASSED test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_simple[cache] PASSED test_backup_restore_on_cluster/test.py::test_shutdown_waits_for_backup PASSED test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_simple[cache] PASSED test_backup_restore_on_cluster/test.py::test_system_functions PASSED test_backup_restore_on_cluster/test.py::test_system_users PASSED test_dictionaries_redis/test.py::test_redis_dictionaries[String_] PASSED test_async_load_databases/test.py::test_dict_get_data PASSED test_backup_restore_on_cluster/test.py::test_table_in_replicated_database_with_not_synced_def PASSED test_backup_restore_on_cluster/test.py::test_table_with_parts_in_queue_considered_non_empty PASSED test_azure_blob_storage_zero_copy_replication/test.py::test_zero_copy_replication PASSED test_backup_restore_on_cluster/test.py::test_tables_dependency PASSED test_async_load_databases/test.py::test_multiple_tables PASSED test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_simple[direct] PASSED test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_simple[direct] PASSED test_config_corresponding_root/test.py::test_work PASSED test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_simple[flat] PASSED test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_simple[flat] PASSED test_dictionaries_all_layouts_separate_sources/test_clickhouse_local.py::test_simple[hashed] PASSED test_dictionaries_all_layouts_separate_sources/test_mysql.py::test_simple[hashed] ======================= 100 passed in 360.09s (0:06:00) ========================